var/home/core/zuul-output/0000755000175000017500000000000015112330372014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112342203015462 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005250462515112342175017707 0ustar rootrootNov 28 14:29:36 crc systemd[1]: Starting Kubernetes Kubelet... Nov 28 14:29:36 crc restorecon[4690]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:36 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 14:29:37 crc restorecon[4690]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 28 14:29:37 crc kubenswrapper[4817]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 14:29:37 crc kubenswrapper[4817]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 28 14:29:37 crc kubenswrapper[4817]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 14:29:37 crc kubenswrapper[4817]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 14:29:37 crc kubenswrapper[4817]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 28 14:29:37 crc kubenswrapper[4817]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.554151 4817 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560757 4817 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560787 4817 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560797 4817 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560806 4817 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560814 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560822 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560830 4817 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560841 4817 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560852 4817 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560861 4817 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560869 4817 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560878 4817 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560886 4817 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560894 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560902 4817 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560910 4817 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560918 4817 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560927 4817 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560934 4817 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560943 4817 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560950 4817 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560958 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560966 4817 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560974 4817 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560982 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560992 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.560999 4817 feature_gate.go:330] unrecognized feature gate: Example Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561007 4817 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561014 4817 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561022 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561030 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561038 4817 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561045 4817 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561055 4817 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561068 4817 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561078 4817 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561087 4817 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561095 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561103 4817 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561112 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561122 4817 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561132 4817 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561139 4817 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561147 4817 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561155 4817 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561163 4817 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561171 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561179 4817 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561187 4817 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561194 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561203 4817 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561210 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561218 4817 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561226 4817 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561234 4817 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561242 4817 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561250 4817 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561258 4817 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561265 4817 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561274 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561285 4817 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561295 4817 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561305 4817 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561313 4817 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561321 4817 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561329 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561337 4817 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561345 4817 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561352 4817 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561360 4817 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.561368 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561753 4817 flags.go:64] FLAG: --address="0.0.0.0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561774 4817 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561794 4817 flags.go:64] FLAG: --anonymous-auth="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561805 4817 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561818 4817 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561828 4817 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561839 4817 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561850 4817 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561860 4817 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561869 4817 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561879 4817 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561889 4817 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561899 4817 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561908 4817 flags.go:64] FLAG: --cgroup-root="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561919 4817 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561929 4817 flags.go:64] FLAG: --client-ca-file="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561938 4817 flags.go:64] FLAG: --cloud-config="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561947 4817 flags.go:64] FLAG: --cloud-provider="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561956 4817 flags.go:64] FLAG: --cluster-dns="[]" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561966 4817 flags.go:64] FLAG: --cluster-domain="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561976 4817 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561985 4817 flags.go:64] FLAG: --config-dir="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.561994 4817 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562003 4817 flags.go:64] FLAG: --container-log-max-files="5" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562014 4817 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562023 4817 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562032 4817 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562041 4817 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562050 4817 flags.go:64] FLAG: --contention-profiling="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562092 4817 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562101 4817 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562111 4817 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562119 4817 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562131 4817 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562140 4817 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562148 4817 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562158 4817 flags.go:64] FLAG: --enable-load-reader="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562167 4817 flags.go:64] FLAG: --enable-server="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562176 4817 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562188 4817 flags.go:64] FLAG: --event-burst="100" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562204 4817 flags.go:64] FLAG: --event-qps="50" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562214 4817 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562223 4817 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562232 4817 flags.go:64] FLAG: --eviction-hard="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562243 4817 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562251 4817 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562261 4817 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562270 4817 flags.go:64] FLAG: --eviction-soft="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562302 4817 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562313 4817 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562322 4817 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562332 4817 flags.go:64] FLAG: --experimental-mounter-path="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562341 4817 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562350 4817 flags.go:64] FLAG: --fail-swap-on="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562358 4817 flags.go:64] FLAG: --feature-gates="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562369 4817 flags.go:64] FLAG: --file-check-frequency="20s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562378 4817 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562387 4817 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562397 4817 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562406 4817 flags.go:64] FLAG: --healthz-port="10248" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562415 4817 flags.go:64] FLAG: --help="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562424 4817 flags.go:64] FLAG: --hostname-override="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562433 4817 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562442 4817 flags.go:64] FLAG: --http-check-frequency="20s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562451 4817 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562460 4817 flags.go:64] FLAG: --image-credential-provider-config="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562469 4817 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562478 4817 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562489 4817 flags.go:64] FLAG: --image-service-endpoint="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562498 4817 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562506 4817 flags.go:64] FLAG: --kube-api-burst="100" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562515 4817 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562525 4817 flags.go:64] FLAG: --kube-api-qps="50" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562534 4817 flags.go:64] FLAG: --kube-reserved="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562543 4817 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562552 4817 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562562 4817 flags.go:64] FLAG: --kubelet-cgroups="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562571 4817 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562581 4817 flags.go:64] FLAG: --lock-file="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562590 4817 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562599 4817 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562608 4817 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562622 4817 flags.go:64] FLAG: --log-json-split-stream="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562633 4817 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562643 4817 flags.go:64] FLAG: --log-text-split-stream="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562651 4817 flags.go:64] FLAG: --logging-format="text" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562660 4817 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562669 4817 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562678 4817 flags.go:64] FLAG: --manifest-url="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562687 4817 flags.go:64] FLAG: --manifest-url-header="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562698 4817 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562707 4817 flags.go:64] FLAG: --max-open-files="1000000" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562741 4817 flags.go:64] FLAG: --max-pods="110" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562751 4817 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562761 4817 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562769 4817 flags.go:64] FLAG: --memory-manager-policy="None" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562778 4817 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562787 4817 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562796 4817 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562806 4817 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562825 4817 flags.go:64] FLAG: --node-status-max-images="50" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562834 4817 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562843 4817 flags.go:64] FLAG: --oom-score-adj="-999" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562852 4817 flags.go:64] FLAG: --pod-cidr="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562861 4817 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562875 4817 flags.go:64] FLAG: --pod-manifest-path="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562883 4817 flags.go:64] FLAG: --pod-max-pids="-1" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562893 4817 flags.go:64] FLAG: --pods-per-core="0" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562902 4817 flags.go:64] FLAG: --port="10250" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562912 4817 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562920 4817 flags.go:64] FLAG: --provider-id="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562930 4817 flags.go:64] FLAG: --qos-reserved="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562939 4817 flags.go:64] FLAG: --read-only-port="10255" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562949 4817 flags.go:64] FLAG: --register-node="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562958 4817 flags.go:64] FLAG: --register-schedulable="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562966 4817 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562981 4817 flags.go:64] FLAG: --registry-burst="10" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562990 4817 flags.go:64] FLAG: --registry-qps="5" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.562998 4817 flags.go:64] FLAG: --reserved-cpus="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563007 4817 flags.go:64] FLAG: --reserved-memory="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563018 4817 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563027 4817 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563036 4817 flags.go:64] FLAG: --rotate-certificates="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563045 4817 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563054 4817 flags.go:64] FLAG: --runonce="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563063 4817 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563072 4817 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563081 4817 flags.go:64] FLAG: --seccomp-default="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563090 4817 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563099 4817 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563108 4817 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563117 4817 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563127 4817 flags.go:64] FLAG: --storage-driver-password="root" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563135 4817 flags.go:64] FLAG: --storage-driver-secure="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563144 4817 flags.go:64] FLAG: --storage-driver-table="stats" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563153 4817 flags.go:64] FLAG: --storage-driver-user="root" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563161 4817 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563171 4817 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563180 4817 flags.go:64] FLAG: --system-cgroups="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563189 4817 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563202 4817 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563211 4817 flags.go:64] FLAG: --tls-cert-file="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563220 4817 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563231 4817 flags.go:64] FLAG: --tls-min-version="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563240 4817 flags.go:64] FLAG: --tls-private-key-file="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563249 4817 flags.go:64] FLAG: --topology-manager-policy="none" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563257 4817 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563266 4817 flags.go:64] FLAG: --topology-manager-scope="container" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563277 4817 flags.go:64] FLAG: --v="2" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563288 4817 flags.go:64] FLAG: --version="false" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563300 4817 flags.go:64] FLAG: --vmodule="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563310 4817 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.563320 4817 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563539 4817 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563550 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563559 4817 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563567 4817 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563575 4817 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563583 4817 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563594 4817 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563604 4817 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563614 4817 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563624 4817 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563633 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563643 4817 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563652 4817 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563662 4817 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563672 4817 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563681 4817 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563690 4817 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563700 4817 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563709 4817 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563740 4817 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563749 4817 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563757 4817 feature_gate.go:330] unrecognized feature gate: Example Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563775 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563783 4817 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563792 4817 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563800 4817 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563809 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563818 4817 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563826 4817 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563835 4817 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563844 4817 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563854 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563862 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563870 4817 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563878 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563888 4817 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563897 4817 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563905 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563913 4817 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563922 4817 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563929 4817 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563937 4817 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563945 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563953 4817 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563960 4817 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563968 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563976 4817 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563984 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563992 4817 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.563999 4817 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564007 4817 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564015 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564023 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564031 4817 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564041 4817 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564049 4817 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564057 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564065 4817 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564072 4817 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564080 4817 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564088 4817 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564095 4817 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564103 4817 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564110 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564118 4817 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564125 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564133 4817 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564142 4817 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564149 4817 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564157 4817 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.564165 4817 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.564262 4817 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.575248 4817 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.575308 4817 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575424 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575439 4817 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575453 4817 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575460 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575466 4817 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575473 4817 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575480 4817 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575488 4817 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575494 4817 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575502 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575509 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575518 4817 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575528 4817 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575536 4817 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575544 4817 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575550 4817 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575556 4817 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575561 4817 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575568 4817 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575573 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575579 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575585 4817 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575591 4817 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575597 4817 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575603 4817 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575609 4817 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575614 4817 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575622 4817 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575628 4817 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575635 4817 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575642 4817 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575647 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575652 4817 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575658 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575667 4817 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575675 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575682 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575688 4817 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575694 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575699 4817 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575705 4817 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575711 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575716 4817 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575749 4817 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575755 4817 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575761 4817 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575767 4817 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575772 4817 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575777 4817 feature_gate.go:330] unrecognized feature gate: Example Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575782 4817 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575788 4817 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575793 4817 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575799 4817 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575804 4817 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575809 4817 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575815 4817 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575822 4817 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575828 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575834 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575839 4817 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575844 4817 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575850 4817 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575855 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575860 4817 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575865 4817 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575870 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575876 4817 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575881 4817 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575886 4817 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575891 4817 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.575897 4817 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.575909 4817 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576085 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576097 4817 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576104 4817 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576110 4817 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576115 4817 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576121 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576127 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576135 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576142 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576149 4817 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576158 4817 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576164 4817 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576169 4817 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576174 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576180 4817 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576185 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576191 4817 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576196 4817 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576201 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576206 4817 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576212 4817 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576217 4817 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576222 4817 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576227 4817 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576233 4817 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576239 4817 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576245 4817 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576251 4817 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576258 4817 feature_gate.go:330] unrecognized feature gate: Example Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576264 4817 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576270 4817 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576276 4817 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576283 4817 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576291 4817 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576298 4817 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576304 4817 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576311 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576316 4817 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576322 4817 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576328 4817 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576334 4817 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576339 4817 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576344 4817 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576350 4817 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576355 4817 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576361 4817 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576366 4817 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576371 4817 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576377 4817 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576382 4817 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576387 4817 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576394 4817 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576400 4817 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576405 4817 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576413 4817 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576420 4817 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576427 4817 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576433 4817 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576439 4817 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576445 4817 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576451 4817 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576457 4817 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576463 4817 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576468 4817 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576474 4817 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576480 4817 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576487 4817 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576493 4817 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576499 4817 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576504 4817 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.576510 4817 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.576520 4817 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.577058 4817 server.go:940] "Client rotation is on, will bootstrap in background" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.580958 4817 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.581069 4817 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.581783 4817 server.go:997] "Starting client certificate rotation" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.581814 4817 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.582382 4817 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-24 11:24:27.862250595 +0000 UTC Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.582509 4817 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 620h54m50.27974629s for next certificate rotation Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.591057 4817 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.593540 4817 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.606999 4817 log.go:25] "Validated CRI v1 runtime API" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.629912 4817 log.go:25] "Validated CRI v1 image API" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.632368 4817 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.636169 4817 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-28-14-24-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.636214 4817 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.663001 4817 manager.go:217] Machine: {Timestamp:2025-11-28 14:29:37.660833015 +0000 UTC m=+0.248811341 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:14944962-313a-40b6-b900-0cbeaaf380e7 BootID:603c0640-9dfd-46d2-9c89-a0eb080f9b75 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e1:19:01 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e1:19:01 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:82:93:96 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:15:82:b0 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:2a:96:03 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e3:85:cf Speed:-1 Mtu:1496} {Name:eth10 MacAddress:5e:05:0a:c8:78:82 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:92:3c:1b:2b:03:09 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.663402 4817 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.663593 4817 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.664097 4817 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.664412 4817 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.664468 4817 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.664852 4817 topology_manager.go:138] "Creating topology manager with none policy" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.664872 4817 container_manager_linux.go:303] "Creating device plugin manager" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.665168 4817 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.665227 4817 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.665535 4817 state_mem.go:36] "Initialized new in-memory state store" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.665676 4817 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.666978 4817 kubelet.go:418] "Attempting to sync node with API server" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.667012 4817 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.667050 4817 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.667072 4817 kubelet.go:324] "Adding apiserver pod source" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.667090 4817 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.669757 4817 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.670448 4817 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.671012 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.671117 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.671367 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.671474 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.671788 4817 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672604 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672647 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672664 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672679 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672702 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672716 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672759 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672783 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672797 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672811 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672854 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.672868 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.673534 4817 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.674430 4817 server.go:1280] "Started kubelet" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.674483 4817 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.674949 4817 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.674967 4817 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.675669 4817 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 28 14:29:37 crc systemd[1]: Started Kubernetes Kubelet. Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.678581 4817 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c320b313b4636 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 14:29:37.674389046 +0000 UTC m=+0.262367352,LastTimestamp:2025-11-28 14:29:37.674389046 +0000 UTC m=+0.262367352,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.679326 4817 server.go:460] "Adding debug handlers to kubelet server" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.680924 4817 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.680993 4817 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.681043 4817 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 22:22:47.07873515 +0000 UTC Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.681202 4817 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.681220 4817 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.681229 4817 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.681287 4817 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.682203 4817 factory.go:55] Registering systemd factory Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.682227 4817 factory.go:221] Registration of the systemd container factory successfully Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.684098 4817 factory.go:153] Registering CRI-O factory Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.684146 4817 factory.go:221] Registration of the crio container factory successfully Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.684201 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.684278 4817 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.684351 4817 factory.go:103] Registering Raw factory Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.684348 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.684369 4817 manager.go:1196] Started watching for new ooms in manager Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.685668 4817 manager.go:319] Starting recovery of all containers Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.685972 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="200ms" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696483 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696561 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696585 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696604 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696624 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696642 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696670 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696687 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696707 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696752 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696793 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696812 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696830 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696852 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696872 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696891 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696910 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696926 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696943 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696961 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696978 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.696995 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697013 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697032 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697052 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697071 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697093 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697115 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697134 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697152 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697200 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697218 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697239 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697256 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697275 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697294 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697313 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697333 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697352 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697370 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697389 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697407 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697425 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697445 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697463 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697482 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697500 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697521 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697541 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697561 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697581 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697601 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697628 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697649 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697669 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697691 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697777 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697799 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697818 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697837 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697855 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697873 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697891 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697908 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697927 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697946 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697964 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.697982 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698000 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698017 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698044 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698062 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698081 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698099 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698121 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698185 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698211 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698236 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698260 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698284 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698304 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698321 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698341 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698360 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698378 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698396 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698414 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698433 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698453 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698471 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698491 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698510 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698529 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698548 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698565 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698582 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698603 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698622 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698640 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698660 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698681 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698699 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698747 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698767 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698795 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698815 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698834 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698855 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698876 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698896 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698916 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698935 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698956 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698975 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.698998 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699017 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699038 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699059 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699079 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699098 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699116 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699133 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699151 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699173 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699191 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699211 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699229 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699247 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699267 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699286 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699305 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699324 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699343 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699361 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699382 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699402 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699420 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699495 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699514 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699532 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699549 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699568 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699591 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699614 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.699634 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700492 4817 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700532 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700554 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700577 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700599 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700617 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700636 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700653 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700672 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700690 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700710 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700762 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700781 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700799 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700820 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700840 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700858 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700877 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700897 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700917 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700935 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700953 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700970 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.700990 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701009 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701027 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701045 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701066 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701086 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701106 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701125 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701146 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701496 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701534 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.701560 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702503 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702584 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702634 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702654 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702673 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702708 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702748 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702783 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702806 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702830 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702862 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702886 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702917 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702941 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702966 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.702998 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703018 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703049 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703071 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703093 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703119 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703157 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703185 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703206 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703231 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703258 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703280 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703313 4817 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703331 4817 reconstruct.go:97] "Volume reconstruction finished" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.703344 4817 reconciler.go:26] "Reconciler: start to sync state" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.716135 4817 manager.go:324] Recovery completed Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.726432 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.729745 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.729791 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.729802 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.731895 4817 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.731927 4817 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.731968 4817 state_mem.go:36] "Initialized new in-memory state store" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.732457 4817 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.735341 4817 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.735586 4817 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.735647 4817 kubelet.go:2335] "Starting kubelet main sync loop" Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.735769 4817 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 28 14:29:37 crc kubenswrapper[4817]: W1128 14:29:37.738875 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.738974 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.742909 4817 policy_none.go:49] "None policy: Start" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.743801 4817 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.743903 4817 state_mem.go:35] "Initializing new in-memory state store" Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.781829 4817 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.808037 4817 manager.go:334] "Starting Device Plugin manager" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.808138 4817 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.808154 4817 server.go:79] "Starting device plugin registration server" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.808688 4817 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.808707 4817 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.809221 4817 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.809479 4817 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.809491 4817 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.822080 4817 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.836863 4817 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.836982 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.838133 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.838172 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.838187 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.838388 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.838698 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.838798 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.839247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.839298 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.839309 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.839512 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.839638 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.839679 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.840038 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.840065 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.840076 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.840959 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841059 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841068 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841097 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841170 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841336 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841399 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.841457 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842177 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842212 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842245 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842365 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842368 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842487 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.842520 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843157 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843192 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843205 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843626 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843688 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843919 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843951 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.843963 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.847371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.847407 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.847428 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.887440 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="400ms" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906362 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906427 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906474 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906553 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906669 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906707 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906748 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906855 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906925 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906967 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.906997 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.907027 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.907049 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.907152 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.907213 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.908911 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.910799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.910858 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.910877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:37 crc kubenswrapper[4817]: I1128 14:29:37.910915 4817 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 14:29:37 crc kubenswrapper[4817]: E1128 14:29:37.911517 4817 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008761 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008838 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008876 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008899 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008922 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008946 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008970 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.008994 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009004 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009081 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009015 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009125 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009146 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009178 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009218 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009216 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009245 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009257 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009281 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009288 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009303 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009333 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009346 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009434 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009413 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009389 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009506 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009524 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009548 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.009332 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.111701 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.113424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.113499 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.113520 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.113564 4817 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 14:29:38 crc kubenswrapper[4817]: E1128 14:29:38.114224 4817 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.176999 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.183897 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: W1128 14:29:38.203615 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-07a39cc67f05a063382de1781cebd08a6b6617ac8fd1d6ed60833c1fb1522332 WatchSource:0}: Error finding container 07a39cc67f05a063382de1781cebd08a6b6617ac8fd1d6ed60833c1fb1522332: Status 404 returned error can't find the container with id 07a39cc67f05a063382de1781cebd08a6b6617ac8fd1d6ed60833c1fb1522332 Nov 28 14:29:38 crc kubenswrapper[4817]: W1128 14:29:38.206226 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b74ebaf5741cb66096138212493bb25593e859a9cff30ef2434c9ca998ed0311 WatchSource:0}: Error finding container b74ebaf5741cb66096138212493bb25593e859a9cff30ef2434c9ca998ed0311: Status 404 returned error can't find the container with id b74ebaf5741cb66096138212493bb25593e859a9cff30ef2434c9ca998ed0311 Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.206524 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.211912 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.217607 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:38 crc kubenswrapper[4817]: W1128 14:29:38.218288 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-968bce150b4eebfd8cd57c8aa1d1fae59e6a402b56c921328ac86b4a9f7b7e08 WatchSource:0}: Error finding container 968bce150b4eebfd8cd57c8aa1d1fae59e6a402b56c921328ac86b4a9f7b7e08: Status 404 returned error can't find the container with id 968bce150b4eebfd8cd57c8aa1d1fae59e6a402b56c921328ac86b4a9f7b7e08 Nov 28 14:29:38 crc kubenswrapper[4817]: W1128 14:29:38.230386 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-d3d16b6c099b51a0bd407b7b09f3d9bcc33ceb7527445adb42186050859ee052 WatchSource:0}: Error finding container d3d16b6c099b51a0bd407b7b09f3d9bcc33ceb7527445adb42186050859ee052: Status 404 returned error can't find the container with id d3d16b6c099b51a0bd407b7b09f3d9bcc33ceb7527445adb42186050859ee052 Nov 28 14:29:38 crc kubenswrapper[4817]: W1128 14:29:38.246402 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-d35ddfa90485437bc75e46a207f93869ce31f726c7a4e570f553708d98543835 WatchSource:0}: Error finding container d35ddfa90485437bc75e46a207f93869ce31f726c7a4e570f553708d98543835: Status 404 returned error can't find the container with id d35ddfa90485437bc75e46a207f93869ce31f726c7a4e570f553708d98543835 Nov 28 14:29:38 crc kubenswrapper[4817]: E1128 14:29:38.288930 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="800ms" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.514303 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.516288 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.516345 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.516355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.516389 4817 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 14:29:38 crc kubenswrapper[4817]: E1128 14:29:38.517000 4817 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.675629 4817 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.681741 4817 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 06:38:57.694187839 +0000 UTC Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.681821 4817 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 16h9m19.012368786s for next certificate rotation Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.744894 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.745053 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d35ddfa90485437bc75e46a207f93869ce31f726c7a4e570f553708d98543835"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.747952 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8" exitCode=0 Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.748046 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.748130 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d3d16b6c099b51a0bd407b7b09f3d9bcc33ceb7527445adb42186050859ee052"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.748322 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.749822 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.749858 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.749868 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.751749 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.752383 4817 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21" exitCode=0 Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.752470 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.753897 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"968bce150b4eebfd8cd57c8aa1d1fae59e6a402b56c921328ac86b4a9f7b7e08"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.752524 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.753947 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.753994 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.753979 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.755600 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.755641 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.755652 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.756158 4817 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="9e50e590d43ee1c647bf22ff2b37fe9f4b14c7b4115161b6245208bde48073c8" exitCode=0 Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.756306 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"9e50e590d43ee1c647bf22ff2b37fe9f4b14c7b4115161b6245208bde48073c8"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.756428 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"07a39cc67f05a063382de1781cebd08a6b6617ac8fd1d6ed60833c1fb1522332"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.756608 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.757655 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.757693 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.757709 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.763413 4817 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697" exitCode=0 Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.763511 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.763660 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b74ebaf5741cb66096138212493bb25593e859a9cff30ef2434c9ca998ed0311"} Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.764035 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.765590 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.765651 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:38 crc kubenswrapper[4817]: I1128 14:29:38.765663 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:39 crc kubenswrapper[4817]: W1128 14:29:39.022361 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:39 crc kubenswrapper[4817]: E1128 14:29:39.022481 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:39 crc kubenswrapper[4817]: W1128 14:29:39.030111 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:39 crc kubenswrapper[4817]: E1128 14:29:39.031444 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:39 crc kubenswrapper[4817]: E1128 14:29:39.089926 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="1.6s" Nov 28 14:29:39 crc kubenswrapper[4817]: W1128 14:29:39.171020 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:39 crc kubenswrapper[4817]: E1128 14:29:39.171137 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:39 crc kubenswrapper[4817]: W1128 14:29:39.171827 4817 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Nov 28 14:29:39 crc kubenswrapper[4817]: E1128 14:29:39.171933 4817 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.317420 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.319503 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.319556 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.319570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.319605 4817 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 14:29:39 crc kubenswrapper[4817]: E1128 14:29:39.320277 4817 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.769207 4817 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c" exitCode=0 Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.769325 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.769542 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.770676 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.770713 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.770738 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.772262 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"933c4885ef9df2b3a4ca84ff7339657b932898469179395b4ba5140a248f9fa7"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.772467 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.774692 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.774747 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.774762 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.779715 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.779855 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.779881 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.780065 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.781363 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.781417 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.781436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.783605 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.783656 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.783671 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.783706 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.784797 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.784850 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.784861 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.787514 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.787579 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.787596 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd"} Nov 28 14:29:39 crc kubenswrapper[4817]: I1128 14:29:39.787609 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87"} Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.794034 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530"} Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.794093 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.796169 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.796202 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.796211 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.797270 4817 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba" exitCode=0 Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.797334 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba"} Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.797359 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.797549 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.798382 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.798420 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.798431 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.799091 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.799141 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.799151 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.920447 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.922038 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.922095 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.922109 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:40 crc kubenswrapper[4817]: I1128 14:29:40.922147 4817 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.806828 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7"} Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.806881 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.806904 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c"} Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.806930 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158"} Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.806963 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247"} Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.806937 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.808373 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.808416 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.808436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.934758 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.935025 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.936697 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.936845 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.936865 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:41 crc kubenswrapper[4817]: I1128 14:29:41.943144 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.816466 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd"} Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.816619 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.816619 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.818332 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.818409 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.818446 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.818340 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.818515 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:42 crc kubenswrapper[4817]: I1128 14:29:42.818545 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.200783 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.201019 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.202975 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.203013 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.203023 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.298091 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.298364 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.298433 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.301225 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.301286 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.301305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.660996 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.663392 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.819263 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.819951 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.820803 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.820877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.820898 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.821685 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.821841 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:43 crc kubenswrapper[4817]: I1128 14:29:43.821860 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:44 crc kubenswrapper[4817]: I1128 14:29:44.822681 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:44 crc kubenswrapper[4817]: I1128 14:29:44.824917 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:44 crc kubenswrapper[4817]: I1128 14:29:44.824983 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:44 crc kubenswrapper[4817]: I1128 14:29:44.825001 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:45 crc kubenswrapper[4817]: I1128 14:29:45.098960 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:45 crc kubenswrapper[4817]: I1128 14:29:45.099213 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:29:45 crc kubenswrapper[4817]: I1128 14:29:45.099276 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:45 crc kubenswrapper[4817]: I1128 14:29:45.101119 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:45 crc kubenswrapper[4817]: I1128 14:29:45.101156 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:45 crc kubenswrapper[4817]: I1128 14:29:45.101170 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.512698 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.512959 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.514393 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.514455 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.514467 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.903773 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.904027 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.905570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.905675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:46 crc kubenswrapper[4817]: I1128 14:29:46.905699 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:47 crc kubenswrapper[4817]: E1128 14:29:47.823159 4817 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 14:29:48 crc kubenswrapper[4817]: I1128 14:29:48.804929 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 28 14:29:48 crc kubenswrapper[4817]: I1128 14:29:48.805184 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:48 crc kubenswrapper[4817]: I1128 14:29:48.806693 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:48 crc kubenswrapper[4817]: I1128 14:29:48.806795 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:48 crc kubenswrapper[4817]: I1128 14:29:48.806816 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.358930 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.359219 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.361438 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.361491 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.361505 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.363870 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.676475 4817 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.837438 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.838878 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.838954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:49 crc kubenswrapper[4817]: I1128 14:29:49.838981 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:50 crc kubenswrapper[4817]: E1128 14:29:50.692036 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Nov 28 14:29:50 crc kubenswrapper[4817]: I1128 14:29:50.815206 4817 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 14:29:50 crc kubenswrapper[4817]: I1128 14:29:50.815305 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 14:29:50 crc kubenswrapper[4817]: I1128 14:29:50.826656 4817 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 14:29:50 crc kubenswrapper[4817]: I1128 14:29:50.826782 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 14:29:52 crc kubenswrapper[4817]: I1128 14:29:52.359676 4817 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 14:29:52 crc kubenswrapper[4817]: I1128 14:29:52.360934 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.310926 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.311313 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.315785 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.315849 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.315862 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.319168 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.848438 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.849899 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.849959 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:53 crc kubenswrapper[4817]: I1128 14:29:53.849977 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.810539 4817 trace.go:236] Trace[1310472413]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 14:29:41.340) (total time: 14469ms): Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[1310472413]: ---"Objects listed" error: 14469ms (14:29:55.810) Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[1310472413]: [14.469796504s] [14.469796504s] END Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.810588 4817 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.811384 4817 trace.go:236] Trace[17857564]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 14:29:41.359) (total time: 14451ms): Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[17857564]: ---"Objects listed" error: 14451ms (14:29:55.811) Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[17857564]: [14.451563193s] [14.451563193s] END Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.811416 4817 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.811668 4817 trace.go:236] Trace[1864417307]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 14:29:41.210) (total time: 14601ms): Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[1864417307]: ---"Objects listed" error: 14601ms (14:29:55.811) Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[1864417307]: [14.601182022s] [14.601182022s] END Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.811753 4817 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.815848 4817 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 28 14:29:55 crc kubenswrapper[4817]: E1128 14:29:55.817139 4817 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.817973 4817 trace.go:236] Trace[978158042]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 14:29:41.796) (total time: 14021ms): Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[978158042]: ---"Objects listed" error: 14021ms (14:29:55.817) Nov 28 14:29:55 crc kubenswrapper[4817]: Trace[978158042]: [14.021354768s] [14.021354768s] END Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.818022 4817 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.855833 4817 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51508->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.855895 4817 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51516->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.855905 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51508->192.168.126.11:17697: read: connection reset by peer" Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.855964 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51516->192.168.126.11:17697: read: connection reset by peer" Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.856506 4817 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 28 14:29:55 crc kubenswrapper[4817]: I1128 14:29:55.856566 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.513798 4817 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.513865 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.676325 4817 apiserver.go:52] "Watching apiserver" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.680363 4817 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.680650 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.681064 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.681146 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.681071 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.681191 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.681182 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.681243 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.682313 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.684624 4817 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.685244 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.685439 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.685631 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.685652 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.686203 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.686297 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.687500 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.687679 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.687752 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.688095 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.688136 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.719171 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720294 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720366 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720410 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720445 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720481 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720519 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720591 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720630 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720671 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720758 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720807 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720842 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720876 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720912 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720900 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720947 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720981 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721019 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721057 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721120 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721154 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721192 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721229 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721274 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721344 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721379 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721527 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721133 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721539 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721562 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721766 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721983 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722009 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722022 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722048 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722306 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722643 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.720923 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722407 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722443 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722428 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.721571 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722850 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723386 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723448 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723494 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723531 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723565 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723598 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723632 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723667 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723701 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723766 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723816 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723872 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723929 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723972 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724010 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724047 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724081 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724117 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724155 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724195 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724230 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724272 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724312 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724348 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724423 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724458 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724491 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724531 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724630 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724665 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722452 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.722561 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.723667 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725182 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724700 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725277 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725340 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725386 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725422 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725455 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725490 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725523 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725557 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725632 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725667 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725701 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725755 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725790 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725824 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725858 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725889 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725929 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725965 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726090 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726128 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726165 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726200 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726237 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726271 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726311 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726345 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726378 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726418 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726460 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726498 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726538 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726573 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726604 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726635 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726666 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726700 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726758 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726798 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726831 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726865 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726951 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726990 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.727284 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.728325 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729011 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729808 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724021 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724238 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724453 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724682 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.724680 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.725423 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.726769 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.727008 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.727189 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.727522 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.727673 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.728061 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.728596 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.728623 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.728678 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729250 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729324 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729624 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729645 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729755 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729838 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.730042 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.730084 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.730514 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.730535 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.730547 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731018 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731070 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731308 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731335 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731712 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731846 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731859 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.731911 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.729952 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732053 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732050 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732090 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732124 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732159 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732166 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732186 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732214 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732239 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732265 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732293 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732320 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732345 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732369 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732394 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732418 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732449 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732476 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732395 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732502 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732526 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732574 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732610 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732607 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732637 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732701 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732740 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732763 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732785 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732810 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732833 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732856 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732884 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732900 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732905 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732962 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.732996 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733122 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733170 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733195 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733214 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733234 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733253 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733271 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733290 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733290 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733310 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733333 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733353 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733375 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733392 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733410 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733428 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733447 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733466 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733558 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733577 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733709 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733744 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733761 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733778 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733794 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733813 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733831 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733852 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733873 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733894 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733914 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733937 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733959 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733981 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733999 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.734017 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.734035 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.734054 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742108 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742510 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742780 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742960 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743117 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743272 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743438 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743576 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743715 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743890 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744045 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744207 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744352 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744487 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744672 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744943 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745218 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745409 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745891 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745805 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.746641 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.746829 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747012 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747181 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747372 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747837 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748043 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748215 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748379 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748532 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748853 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749046 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749235 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749405 4817 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749558 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749758 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750054 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750210 4817 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750363 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750508 4817 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750633 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.751282 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.751336 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.752200 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.752511 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.752686 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.751172 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.733345 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.753424 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.735091 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.735536 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.736155 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.736289 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.736880 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.736898 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.737538 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.754466 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.737541 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.754505 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.737767 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.737857 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.737878 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738046 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738065 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738282 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738281 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738294 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738389 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738352 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738397 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738449 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738479 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.754658 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738573 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738701 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.738975 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.739177 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.739199 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.765419 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.739522 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.764664 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.739877 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.739923 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.740799 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.741356 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.741413 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.741508 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.766703 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.741567 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742043 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742779 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742801 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742846 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742898 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.742928 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.743912 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.744043 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745509 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745564 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.745714 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747878 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.747982 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748142 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748365 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.748664 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749244 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749492 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749533 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749569 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749681 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749684 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749954 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.749326 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750385 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750534 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750767 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.750216 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.751040 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.751133 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.751247 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.751240 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.751631 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.753528 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.753668 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.753844 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.754396 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.754489 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.755131 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.755162 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.755249 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.755628 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.755769 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.755878 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.756299 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.757532 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.760071 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.765650 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.765861 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.769035 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.757968 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.769472 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.769495 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.769513 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.770071 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.753178 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:29:57.253138676 +0000 UTC m=+19.841117002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.772026 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:57.271990083 +0000 UTC m=+19.859968359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.772072 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:57.272058955 +0000 UTC m=+19.860037341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.772155 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:57.272141807 +0000 UTC m=+19.860120083 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.771427 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772247 4817 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772276 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772296 4817 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772316 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772331 4817 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772347 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772363 4817 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772379 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772397 4817 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772414 4817 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772431 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772449 4817 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772464 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772479 4817 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772495 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772509 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772523 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772538 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772552 4817 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772565 4817 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772665 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772683 4817 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772697 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772711 4817 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772745 4817 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772759 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772775 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772789 4817 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772807 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772827 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772842 4817 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772857 4817 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772869 4817 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772886 4817 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772904 4817 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772920 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772934 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772951 4817 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772967 4817 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772984 4817 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.772999 4817 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.773014 4817 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.773031 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.773048 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.773061 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.773076 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.773091 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.774700 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.776864 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.778456 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.778686 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.778737 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.778753 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:56 crc kubenswrapper[4817]: E1128 14:29:56.778813 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:57.278789426 +0000 UTC m=+19.866767702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.779057 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.779076 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.779279 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.779483 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.779493 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.779599 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.780158 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.782111 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.782222 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.782049 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.782676 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.783197 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.783204 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.783297 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.783231 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.783397 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.783511 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.784298 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.784339 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.784668 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.785368 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.788228 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.788393 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.788987 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.791052 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.791835 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.791854 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.795081 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.795118 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.795634 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.797951 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.798289 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.798418 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.798540 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.798713 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.799474 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.799613 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.801609 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.801654 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.803672 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.803994 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.805066 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.811629 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.814963 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.817960 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.835068 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.836677 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.854222 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.858482 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.864176 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.867605 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530" exitCode=255 Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.867652 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530"} Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.874794 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.874890 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875074 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875336 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875495 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875638 4817 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875675 4817 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875649 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875758 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875770 4817 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875781 4817 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875793 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875802 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875813 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875825 4817 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875835 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875845 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875855 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875865 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875876 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875886 4817 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875895 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875906 4817 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875917 4817 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875926 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875936 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875946 4817 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875957 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875968 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875979 4817 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.875989 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876000 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876010 4817 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876021 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876031 4817 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876043 4817 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876055 4817 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876065 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876076 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876087 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876097 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876108 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876119 4817 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876129 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876140 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876151 4817 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876173 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876184 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876195 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876208 4817 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876218 4817 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876229 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876239 4817 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876249 4817 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876259 4817 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876269 4817 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876282 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876293 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876305 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876318 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876329 4817 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876344 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876355 4817 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876368 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876379 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876390 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876401 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876411 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876421 4817 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876431 4817 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876442 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876453 4817 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876463 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876476 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876488 4817 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876499 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876511 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876522 4817 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876532 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876541 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876552 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876561 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876572 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876581 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876589 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876598 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876607 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876616 4817 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876649 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876658 4817 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876667 4817 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876677 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876686 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876695 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876704 4817 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876713 4817 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876735 4817 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876744 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876755 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876765 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876776 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876787 4817 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876795 4817 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876804 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876813 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876822 4817 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876830 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876839 4817 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876848 4817 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876858 4817 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876867 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876876 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876886 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876895 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876904 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876913 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876922 4817 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876931 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876943 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876955 4817 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876968 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876976 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876985 4817 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.876996 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877005 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877014 4817 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877023 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877032 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877045 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877054 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877063 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877072 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877082 4817 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877090 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877099 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877108 4817 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877117 4817 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877124 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877133 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877142 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877151 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.877159 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.883486 4817 scope.go:117] "RemoveContainer" containerID="dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.883922 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.884197 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.898109 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.912677 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.921908 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.941610 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.980171 4817 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 14:29:56 crc kubenswrapper[4817]: I1128 14:29:56.994987 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.004634 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.018361 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 14:29:57 crc kubenswrapper[4817]: W1128 14:29:57.022208 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-4218f75ef4d07f7be8a0a5da93e3d11966618987e8741419a42d96f577da6e07 WatchSource:0}: Error finding container 4218f75ef4d07f7be8a0a5da93e3d11966618987e8741419a42d96f577da6e07: Status 404 returned error can't find the container with id 4218f75ef4d07f7be8a0a5da93e3d11966618987e8741419a42d96f577da6e07 Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.039074 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.282193 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.282278 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.282314 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.282345 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.282370 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282409 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:29:58.282382758 +0000 UTC m=+20.870361024 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282497 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282548 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282560 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:58.282542143 +0000 UTC m=+20.870520409 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282564 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282588 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282611 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282622 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:58.282612245 +0000 UTC m=+20.870590511 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282623 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282644 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282659 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282678 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:58.282671666 +0000 UTC m=+20.870649932 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.282694 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:29:58.282687797 +0000 UTC m=+20.870666063 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.736071 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:29:57 crc kubenswrapper[4817]: E1128 14:29:57.736223 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.739585 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.740475 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.741392 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.742187 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.742933 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.743560 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.745634 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.746336 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.747664 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.748424 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.749572 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.750372 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.751176 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.751635 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.753036 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.753564 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.754542 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.757593 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.758214 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.759038 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.760771 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.761335 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.762218 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.762855 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.763278 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.764468 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.765566 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.766072 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.766623 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.767431 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.767962 4817 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.768057 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.771060 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.771582 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.772011 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.773022 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.773566 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.774923 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.775502 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.776509 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.777297 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.778413 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.780715 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.782108 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.782869 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.783604 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.784397 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.785146 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.786905 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.787553 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.788801 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.789302 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.790435 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.791047 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.791541 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.798754 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.829441 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.844830 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.854705 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-wn7r7"] Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.855074 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.856308 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.856527 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.857022 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.860134 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.876905 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.876934 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.880513 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.880877 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.881367 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9d8c5f56966181db5fbbc778a5ffb3b7013a5c68f04bd15308c7cff5dfef6853"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.883593 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.883637 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.883652 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d5e63b939b35fbf9c5ecdee037334cfa1c6302413eb57ea17d67ae5e0a272e6c"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.890189 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.891797 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.891862 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4218f75ef4d07f7be8a0a5da93e3d11966618987e8741419a42d96f577da6e07"} Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.903049 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.918578 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.928209 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.939964 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.950246 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.963838 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.979648 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.987904 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fff6f26f-fe95-4e3f-a7d8-123897ca4bf6-hosts-file\") pod \"node-resolver-wn7r7\" (UID: \"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\") " pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.987984 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bhts\" (UniqueName: \"kubernetes.io/projected/fff6f26f-fe95-4e3f-a7d8-123897ca4bf6-kube-api-access-8bhts\") pod \"node-resolver-wn7r7\" (UID: \"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\") " pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:57 crc kubenswrapper[4817]: I1128 14:29:57.998683 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.013841 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.033463 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.047349 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.062295 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.074704 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.089552 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.089847 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fff6f26f-fe95-4e3f-a7d8-123897ca4bf6-hosts-file\") pod \"node-resolver-wn7r7\" (UID: \"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\") " pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.089913 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bhts\" (UniqueName: \"kubernetes.io/projected/fff6f26f-fe95-4e3f-a7d8-123897ca4bf6-kube-api-access-8bhts\") pod \"node-resolver-wn7r7\" (UID: \"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\") " pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.090015 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fff6f26f-fe95-4e3f-a7d8-123897ca4bf6-hosts-file\") pod \"node-resolver-wn7r7\" (UID: \"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\") " pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.111029 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bhts\" (UniqueName: \"kubernetes.io/projected/fff6f26f-fe95-4e3f-a7d8-123897ca4bf6-kube-api-access-8bhts\") pod \"node-resolver-wn7r7\" (UID: \"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\") " pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.111470 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.130576 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.165694 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wn7r7" Nov 28 14:29:58 crc kubenswrapper[4817]: W1128 14:29:58.184885 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfff6f26f_fe95_4e3f_a7d8_123897ca4bf6.slice/crio-1433aa3745c3983646a5e0d2baa2e91a524eded4adaa889a55b5e22fe7a92f60 WatchSource:0}: Error finding container 1433aa3745c3983646a5e0d2baa2e91a524eded4adaa889a55b5e22fe7a92f60: Status 404 returned error can't find the container with id 1433aa3745c3983646a5e0d2baa2e91a524eded4adaa889a55b5e22fe7a92f60 Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.219997 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bktf8"] Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.221035 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rx9ch"] Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.221225 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.222399 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-6kh49"] Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.222649 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.223462 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-sbhgx"] Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.223823 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.224048 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.227291 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.227417 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.227769 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.227906 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228190 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228195 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228324 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228462 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228489 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228498 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228505 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228544 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228590 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228603 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.228677 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.230547 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.230746 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.232731 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.233095 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.250923 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.275107 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.291438 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.291541 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291578 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:30:00.291558276 +0000 UTC m=+22.879536542 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.291607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.291633 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.291653 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291734 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291747 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291768 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291785 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291789 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291770 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:00.291762141 +0000 UTC m=+22.879740407 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291841 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291859 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291864 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:00.291850714 +0000 UTC m=+22.879828980 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291873 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291909 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:00.291877564 +0000 UTC m=+22.879855840 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.291935 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:00.291925076 +0000 UTC m=+22.879903472 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.311526 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.357960 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.384330 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392083 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvq6d\" (UniqueName: \"kubernetes.io/projected/8fa7491e-b213-4068-b2a8-77bf4979fc38-kube-api-access-qvq6d\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392144 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-socket-dir-parent\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392169 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-slash\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392197 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-log-socket\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392219 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-script-lib\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392318 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-netd\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392397 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392429 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-proxy-tls\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392476 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-kubelet\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392507 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-var-lib-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392534 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-etc-kubernetes\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392589 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392616 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-os-release\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392659 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-systemd-units\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392733 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-cnibin\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392770 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e19f2800-93c0-4382-8a63-483f11e16632-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392816 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-cni-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392839 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-multus-certs\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392883 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-mcd-auth-proxy-config\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392917 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-hostroot\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392951 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-ovn-kubernetes\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.392989 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovn-node-metrics-cert\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393013 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-kubelet\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393035 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-node-log\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393067 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-os-release\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393097 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-rootfs\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393125 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-conf-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393149 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9g5\" (UniqueName: \"kubernetes.io/projected/89086178-4127-42a3-8454-68e6d00c912a-kube-api-access-sf9g5\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393195 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-cni-multus\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393220 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-ovn\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393244 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-env-overrides\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393268 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-cnibin\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393294 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89086178-4127-42a3-8454-68e6d00c912a-cni-binary-copy\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393320 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-netns\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393347 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-systemd\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393372 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-etc-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393406 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-config\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393469 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-system-cni-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393495 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89086178-4127-42a3-8454-68e6d00c912a-multus-daemon-config\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393518 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-system-cni-dir\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393548 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e19f2800-93c0-4382-8a63-483f11e16632-cni-binary-copy\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393580 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsw76\" (UniqueName: \"kubernetes.io/projected/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-kube-api-access-jsw76\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393614 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-cni-bin\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393637 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-bin\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393667 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtdv7\" (UniqueName: \"kubernetes.io/projected/e19f2800-93c0-4382-8a63-483f11e16632-kube-api-access-qtdv7\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393697 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-k8s-cni-cncf-io\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393743 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-netns\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.393769 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.413107 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.428862 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.446141 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.459773 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.473027 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.487364 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495280 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-kubelet\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495313 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-node-log\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495332 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-os-release\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495340 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-kubelet\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495349 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-rootfs\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495371 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-rootfs\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495398 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-conf-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495418 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9g5\" (UniqueName: \"kubernetes.io/projected/89086178-4127-42a3-8454-68e6d00c912a-kube-api-access-sf9g5\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495414 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-node-log\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495455 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-cni-multus\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495470 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-conf-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495437 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-cni-multus\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495570 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-ovn\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495596 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-env-overrides\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495627 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-cnibin\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495655 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-ovn\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495663 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89086178-4127-42a3-8454-68e6d00c912a-cni-binary-copy\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495740 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-netns\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495773 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-systemd\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495790 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-cnibin\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495794 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-etc-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495818 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-netns\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495825 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-systemd\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495820 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-config\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495857 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-etc-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495865 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-system-cni-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495883 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89086178-4127-42a3-8454-68e6d00c912a-multus-daemon-config\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495899 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-system-cni-dir\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495913 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e19f2800-93c0-4382-8a63-483f11e16632-cni-binary-copy\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495929 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsw76\" (UniqueName: \"kubernetes.io/projected/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-kube-api-access-jsw76\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495955 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-cni-bin\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495972 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-bin\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.495989 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtdv7\" (UniqueName: \"kubernetes.io/projected/e19f2800-93c0-4382-8a63-483f11e16632-kube-api-access-qtdv7\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496007 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-k8s-cni-cncf-io\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496020 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-netns\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496037 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496052 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvq6d\" (UniqueName: \"kubernetes.io/projected/8fa7491e-b213-4068-b2a8-77bf4979fc38-kube-api-access-qvq6d\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496073 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-socket-dir-parent\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496093 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-slash\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496113 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-log-socket\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496137 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-script-lib\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496157 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-netd\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496177 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496196 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-proxy-tls\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496214 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-kubelet\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496233 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-var-lib-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496252 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-etc-kubernetes\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496273 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496312 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-os-release\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496332 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-systemd-units\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496353 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-socket-dir-parent\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496361 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-cnibin\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496377 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e19f2800-93c0-4382-8a63-483f11e16632-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496384 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-system-cni-dir\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496423 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-cni-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496441 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-multus-certs\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496456 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-mcd-auth-proxy-config\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496473 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-hostroot\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496489 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-ovn-kubernetes\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496551 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovn-node-metrics-cert\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496559 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89086178-4127-42a3-8454-68e6d00c912a-cni-binary-copy\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496619 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496664 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-slash\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496692 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-config\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496697 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-log-socket\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496712 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89086178-4127-42a3-8454-68e6d00c912a-multus-daemon-config\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496775 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-system-cni-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496813 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-netd\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496949 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-os-release\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496998 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-var-lib-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497017 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-os-release\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497026 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e19f2800-93c0-4382-8a63-483f11e16632-cni-binary-copy\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497035 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-kubelet\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.496311 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-env-overrides\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497103 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-systemd-units\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497129 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-bin\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497128 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-etc-kubernetes\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497158 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-openvswitch\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497147 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-cnibin\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497192 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-var-lib-cni-bin\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497203 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-k8s-cni-cncf-io\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497222 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-netns\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497324 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-host-run-multus-certs\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497356 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-hostroot\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497393 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-ovn-kubernetes\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497416 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89086178-4127-42a3-8454-68e6d00c912a-multus-cni-dir\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497836 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e19f2800-93c0-4382-8a63-483f11e16632-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.497907 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-mcd-auth-proxy-config\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.498000 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e19f2800-93c0-4382-8a63-483f11e16632-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.498281 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-script-lib\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.501166 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovn-node-metrics-cert\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.502452 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-proxy-tls\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.502928 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.515655 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.518474 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsw76\" (UniqueName: \"kubernetes.io/projected/2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc-kube-api-access-jsw76\") pod \"machine-config-daemon-6kh49\" (UID: \"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\") " pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.521481 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9g5\" (UniqueName: \"kubernetes.io/projected/89086178-4127-42a3-8454-68e6d00c912a-kube-api-access-sf9g5\") pod \"multus-sbhgx\" (UID: \"89086178-4127-42a3-8454-68e6d00c912a\") " pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.522059 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtdv7\" (UniqueName: \"kubernetes.io/projected/e19f2800-93c0-4382-8a63-483f11e16632-kube-api-access-qtdv7\") pod \"multus-additional-cni-plugins-rx9ch\" (UID: \"e19f2800-93c0-4382-8a63-483f11e16632\") " pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.526473 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvq6d\" (UniqueName: \"kubernetes.io/projected/8fa7491e-b213-4068-b2a8-77bf4979fc38-kube-api-access-qvq6d\") pod \"ovnkube-node-bktf8\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.527833 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.550513 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.550615 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.561345 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.563949 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.568221 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.576765 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sbhgx" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.592404 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.613322 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.632295 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: W1128 14:29:58.634520 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fa7491e_b213_4068_b2a8_77bf4979fc38.slice/crio-07f7ea85e9aafc4c3165b2709de8c98e364135b61299b942e1516da44313ad77 WatchSource:0}: Error finding container 07f7ea85e9aafc4c3165b2709de8c98e364135b61299b942e1516da44313ad77: Status 404 returned error can't find the container with id 07f7ea85e9aafc4c3165b2709de8c98e364135b61299b942e1516da44313ad77 Nov 28 14:29:58 crc kubenswrapper[4817]: W1128 14:29:58.636663 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f2c731e_6fa9_4e0c_81dd_115f13fa4dfc.slice/crio-8efc4d177b104c92774619ad80735f09c044c0c4e9610c212c7c140987e78b1e WatchSource:0}: Error finding container 8efc4d177b104c92774619ad80735f09c044c0c4e9610c212c7c140987e78b1e: Status 404 returned error can't find the container with id 8efc4d177b104c92774619ad80735f09c044c0c4e9610c212c7c140987e78b1e Nov 28 14:29:58 crc kubenswrapper[4817]: W1128 14:29:58.637598 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode19f2800_93c0_4382_8a63_483f11e16632.slice/crio-30874ed19cc4b05bb797a9b2c4be208da2aa0680b9fd0eac25d417d5474505bc WatchSource:0}: Error finding container 30874ed19cc4b05bb797a9b2c4be208da2aa0680b9fd0eac25d417d5474505bc: Status 404 returned error can't find the container with id 30874ed19cc4b05bb797a9b2c4be208da2aa0680b9fd0eac25d417d5474505bc Nov 28 14:29:58 crc kubenswrapper[4817]: W1128 14:29:58.639141 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89086178_4127_42a3_8454_68e6d00c912a.slice/crio-26483b71bdbe5958578af93f2c5ec07ea9aef85e150ca35fdacbb97a1a3f4044 WatchSource:0}: Error finding container 26483b71bdbe5958578af93f2c5ec07ea9aef85e150ca35fdacbb97a1a3f4044: Status 404 returned error can't find the container with id 26483b71bdbe5958578af93f2c5ec07ea9aef85e150ca35fdacbb97a1a3f4044 Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.650587 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.667550 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.735911 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.735979 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.736054 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.736140 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.835965 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.849137 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.849551 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.861142 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.867436 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.893287 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.895884 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerStarted","Data":"e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.895940 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerStarted","Data":"26483b71bdbe5958578af93f2c5ec07ea9aef85e150ca35fdacbb97a1a3f4044"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.897655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.897688 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"8efc4d177b104c92774619ad80735f09c044c0c4e9610c212c7c140987e78b1e"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.898976 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wn7r7" event={"ID":"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6","Type":"ContainerStarted","Data":"fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.899016 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wn7r7" event={"ID":"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6","Type":"ContainerStarted","Data":"1433aa3745c3983646a5e0d2baa2e91a524eded4adaa889a55b5e22fe7a92f60"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.900223 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" exitCode=0 Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.900294 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.900350 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"07f7ea85e9aafc4c3165b2709de8c98e364135b61299b942e1516da44313ad77"} Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.901189 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerStarted","Data":"30874ed19cc4b05bb797a9b2c4be208da2aa0680b9fd0eac25d417d5474505bc"} Nov 28 14:29:58 crc kubenswrapper[4817]: E1128 14:29:58.932928 4817 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.954560 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:58 crc kubenswrapper[4817]: I1128 14:29:58.994384 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.017336 4817 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.021650 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.021686 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.021694 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.021857 4817 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.041252 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.052949 4817 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.053184 4817 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.054034 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.054053 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.054062 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.054077 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.054087 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.079995 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.086048 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.089482 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.089517 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.089527 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.089543 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.089553 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.094168 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.105493 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.109590 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.109634 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.109646 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.109664 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.109674 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.110166 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.120383 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.122814 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.123063 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.123097 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.123107 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.123122 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.123130 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.133469 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.134842 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.141269 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.141305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.141314 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.141340 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.141351 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.151387 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.157944 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.158064 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.160512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.160569 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.160579 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.160599 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.160609 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.167978 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.178333 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.192640 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.208795 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.228589 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.243895 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.261078 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.262678 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.262862 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.262954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.263055 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.263142 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.273056 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.285076 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.297123 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.315837 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.333870 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.352948 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.362761 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.366233 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.366417 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.366512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.366599 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.366670 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.371048 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.372489 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.380228 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.394038 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.404816 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.419289 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.430141 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.440320 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.450838 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.469428 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.469459 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.469469 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.469487 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.469499 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.473069 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.513838 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.546740 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.571679 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.571712 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.571740 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.571759 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.571770 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.587007 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.635183 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.674310 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.674622 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.674644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.674654 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.674672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.674683 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.710097 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.736263 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.736406 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.753545 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.777244 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.777314 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.777334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.777354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.777368 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.801273 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.827328 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.869329 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.880272 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.880313 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.880325 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.880366 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.880381 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.906201 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.908815 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.909409 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.909480 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.909495 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.909506 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.910906 4817 generic.go:334] "Generic (PLEG): container finished" podID="e19f2800-93c0-4382-8a63-483f11e16632" containerID="f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f" exitCode=0 Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.910947 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerDied","Data":"f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f"} Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.915774 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19"} Nov 28 14:29:59 crc kubenswrapper[4817]: E1128 14:29:59.943084 4817 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.966988 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:29:59Z is after 2025-08-24T17:21:41Z" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.984656 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.984711 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.984745 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.984763 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:29:59 crc kubenswrapper[4817]: I1128 14:29:59.984774 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:29:59Z","lastTransitionTime":"2025-11-28T14:29:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.007974 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.048150 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.087204 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.087258 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.087274 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.087297 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.087313 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.094547 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.127156 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.167969 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.191848 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.191893 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.191908 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.191929 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.191944 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.204707 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.246298 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.286351 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.294269 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.294319 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.294337 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.294361 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.294378 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.317511 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.317576 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.317597 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.317622 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.317649 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317710 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:30:04.317695562 +0000 UTC m=+26.905673828 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317780 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317796 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317795 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317811 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317827 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317825 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317847 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317881 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:04.317863106 +0000 UTC m=+26.905841392 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317906 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:04.317894867 +0000 UTC m=+26.905873143 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.317927 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:04.317916288 +0000 UTC m=+26.905894564 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.318031 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.318087 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:04.318077322 +0000 UTC m=+26.906055598 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.325038 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.363861 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.397406 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.397444 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.397453 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.397470 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.397478 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.412773 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.446452 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.487500 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.499606 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.499636 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.499649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.499667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.499678 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.527862 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.587973 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.602832 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.602921 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.602940 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.602969 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.602989 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.615261 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.653790 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.693944 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.705707 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.705766 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.705778 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.705800 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.705811 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.728972 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.736298 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.736380 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.736442 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:00 crc kubenswrapper[4817]: E1128 14:30:00.736569 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.770168 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.808055 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.808108 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.808119 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.808140 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.808153 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.814919 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.910247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.910305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.910317 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.910337 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.910351 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:00Z","lastTransitionTime":"2025-11-28T14:30:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.924064 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.924122 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.925941 4817 generic.go:334] "Generic (PLEG): container finished" podID="e19f2800-93c0-4382-8a63-483f11e16632" containerID="553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32" exitCode=0 Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.925995 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerDied","Data":"553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32"} Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.940633 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.956992 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.971694 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:00 crc kubenswrapper[4817]: I1128 14:30:00.993655 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.025447 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.025490 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.025505 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.025531 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.025546 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.032744 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.061824 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.087680 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.125198 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.128529 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.128574 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.128586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.128607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.128619 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.164839 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.214162 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.231519 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.231582 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.231592 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.231612 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.231628 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.247461 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.272402 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5nszl"] Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.272889 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.286068 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.298024 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.319139 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.326637 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efa9319a-a051-418e-9d47-7720827d45fb-host\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.326707 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/efa9319a-a051-418e-9d47-7720827d45fb-serviceca\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.326818 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88mn9\" (UniqueName: \"kubernetes.io/projected/efa9319a-a051-418e-9d47-7720827d45fb-kube-api-access-88mn9\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.335280 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.335327 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.335341 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.335367 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.335378 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.338126 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.357498 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.411923 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.428273 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efa9319a-a051-418e-9d47-7720827d45fb-host\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.428333 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/efa9319a-a051-418e-9d47-7720827d45fb-serviceca\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.428374 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88mn9\" (UniqueName: \"kubernetes.io/projected/efa9319a-a051-418e-9d47-7720827d45fb-kube-api-access-88mn9\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.428449 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efa9319a-a051-418e-9d47-7720827d45fb-host\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.429638 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/efa9319a-a051-418e-9d47-7720827d45fb-serviceca\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.437987 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.438025 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.438036 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.438056 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.438070 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.449760 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.478369 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88mn9\" (UniqueName: \"kubernetes.io/projected/efa9319a-a051-418e-9d47-7720827d45fb-kube-api-access-88mn9\") pod \"node-ca-5nszl\" (UID: \"efa9319a-a051-418e-9d47-7720827d45fb\") " pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.507542 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.541635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.541687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.541699 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.541741 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.541759 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.556398 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.590888 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5nszl" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.594466 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: W1128 14:30:01.613326 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefa9319a_a051_418e_9d47_7720827d45fb.slice/crio-180a9a4a2dc7bbe7562badf97c51530efcbf00e1e23eff3f0b624ffc11737b44 WatchSource:0}: Error finding container 180a9a4a2dc7bbe7562badf97c51530efcbf00e1e23eff3f0b624ffc11737b44: Status 404 returned error can't find the container with id 180a9a4a2dc7bbe7562badf97c51530efcbf00e1e23eff3f0b624ffc11737b44 Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.627167 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.645508 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.645562 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.645574 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.645593 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.645606 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.671553 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.710939 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.736541 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:01 crc kubenswrapper[4817]: E1128 14:30:01.736780 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.749574 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.751690 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.751797 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.751815 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.751852 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.751871 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.787757 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.834209 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.854502 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.854531 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.854540 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.854559 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.854569 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.868472 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.907808 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.932005 4817 generic.go:334] "Generic (PLEG): container finished" podID="e19f2800-93c0-4382-8a63-483f11e16632" containerID="e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af" exitCode=0 Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.932057 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerDied","Data":"e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.935046 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5nszl" event={"ID":"efa9319a-a051-418e-9d47-7720827d45fb","Type":"ContainerStarted","Data":"dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.935095 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5nszl" event={"ID":"efa9319a-a051-418e-9d47-7720827d45fb","Type":"ContainerStarted","Data":"180a9a4a2dc7bbe7562badf97c51530efcbf00e1e23eff3f0b624ffc11737b44"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.950573 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.957887 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.957948 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.957966 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.958005 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.958023 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:01Z","lastTransitionTime":"2025-11-28T14:30:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:01 crc kubenswrapper[4817]: I1128 14:30:01.991498 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:01Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.026759 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.060543 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.060585 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.060594 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.060612 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.060623 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.065918 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.104156 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.152208 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.166848 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.166893 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.166904 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.166922 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.166933 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.187082 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.225458 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.267037 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.269310 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.269371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.269396 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.269426 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.269449 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.322254 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.351499 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.372392 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.372444 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.372456 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.372475 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.372488 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.389851 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.429032 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.467882 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.477132 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.477193 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.477208 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.477227 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.477239 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.514753 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.547111 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.580397 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.580449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.580464 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.580487 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.580502 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.596611 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.634017 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.674499 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.684209 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.684262 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.684280 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.684306 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.684325 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.736855 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:02 crc kubenswrapper[4817]: E1128 14:30:02.737055 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.736862 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:02 crc kubenswrapper[4817]: E1128 14:30:02.737230 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.788030 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.788091 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.788113 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.788145 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.788167 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.891997 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.892414 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.892626 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.892847 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.893018 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.944377 4817 generic.go:334] "Generic (PLEG): container finished" podID="e19f2800-93c0-4382-8a63-483f11e16632" containerID="d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7" exitCode=0 Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.944462 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerDied","Data":"d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.953030 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.973934 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.996280 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.996322 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.996333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.996354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:02 crc kubenswrapper[4817]: I1128 14:30:02.996370 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:02Z","lastTransitionTime":"2025-11-28T14:30:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.001154 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:02Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.030288 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.047639 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.064168 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.077017 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.089820 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.099330 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.099370 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.099379 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.099396 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.099406 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.115249 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.129419 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.148924 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.169232 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.181772 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.195474 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.201769 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.201804 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.201813 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.201828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.201839 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.225528 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.269872 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.304371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.304406 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.304414 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.304431 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.304441 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.407011 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.407044 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.407052 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.407066 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.407075 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.511526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.511593 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.511619 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.511650 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.511673 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.615798 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.615863 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.615880 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.615921 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.615939 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.718368 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.718410 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.718421 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.718437 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.718448 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.735929 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:03 crc kubenswrapper[4817]: E1128 14:30:03.736126 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.820871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.820918 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.820929 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.820944 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.820953 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.923678 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.923782 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.923807 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.923839 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.923863 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:03Z","lastTransitionTime":"2025-11-28T14:30:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.961816 4817 generic.go:334] "Generic (PLEG): container finished" podID="e19f2800-93c0-4382-8a63-483f11e16632" containerID="498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248" exitCode=0 Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.961865 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerDied","Data":"498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248"} Nov 28 14:30:03 crc kubenswrapper[4817]: I1128 14:30:03.984176 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:03Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.008190 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.026468 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.026516 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.026527 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.026544 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.026555 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.028091 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.041781 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.054086 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.071266 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.095824 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.110244 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.122454 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.128279 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.128369 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.128390 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.128417 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.128477 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.144064 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.161915 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.172424 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.184041 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.195076 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.207351 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.231270 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.231341 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.231353 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.231375 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.231389 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.335469 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.335524 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.335541 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.335566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.335583 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.349438 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.349558 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.349606 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.349652 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.349700 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349842 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349882 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349888 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:30:12.34981261 +0000 UTC m=+34.937790936 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349917 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349948 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349960 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349970 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.349899 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.350045 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:12.350013486 +0000 UTC m=+34.937991942 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.350084 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:12.350058607 +0000 UTC m=+34.938036903 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.350114 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:12.350099518 +0000 UTC m=+34.938077814 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.350172 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.350265 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:12.350242812 +0000 UTC m=+34.938221228 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.438796 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.439436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.439448 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.439465 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.439475 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.542308 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.542352 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.542361 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.542378 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.542390 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.645486 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.645554 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.645577 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.645613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.645636 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.735982 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.736150 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.735986 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:04 crc kubenswrapper[4817]: E1128 14:30:04.736268 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.748182 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.748237 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.748253 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.748278 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.748294 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.850793 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.850839 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.850850 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.850869 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.850880 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.953357 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.953434 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.953454 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.953483 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.953502 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:04Z","lastTransitionTime":"2025-11-28T14:30:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.971191 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.971618 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.971690 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.977799 4817 generic.go:334] "Generic (PLEG): container finished" podID="e19f2800-93c0-4382-8a63-483f11e16632" containerID="75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea" exitCode=0 Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.977861 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerDied","Data":"75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea"} Nov 28 14:30:04 crc kubenswrapper[4817]: I1128 14:30:04.998981 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:04Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.014164 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.015044 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.017159 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.040568 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.056055 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.056904 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.056934 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.056970 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.056994 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.057007 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.073591 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.101521 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.115478 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.146152 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.162266 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.162311 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.162322 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.162342 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.162353 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.174578 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.191053 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.206704 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.216527 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.233060 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.247535 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.264961 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.265022 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.265035 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.265090 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.265106 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.280192 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.294372 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.309843 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.319020 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.327901 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.344326 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.355112 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.365241 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.367216 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.367259 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.367272 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.367289 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.367300 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.382563 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.397031 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.459427 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.469479 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.469534 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.469544 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.469563 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.469576 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.475259 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.490685 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.504185 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.518855 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.534691 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:05Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.572397 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.572448 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.572463 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.572490 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.572509 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.675172 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.675231 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.675242 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.675262 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.675276 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.736251 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:05 crc kubenswrapper[4817]: E1128 14:30:05.736427 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.778852 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.778948 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.778967 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.778998 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.779015 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.882245 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.882340 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.882359 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.882388 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.882408 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.985375 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.985479 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.985497 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.985524 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.985547 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:05Z","lastTransitionTime":"2025-11-28T14:30:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.989664 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" event={"ID":"e19f2800-93c0-4382-8a63-483f11e16632","Type":"ContainerStarted","Data":"a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9"} Nov 28 14:30:05 crc kubenswrapper[4817]: I1128 14:30:05.989711 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.024176 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.040158 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.057817 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.076588 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.093345 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.094269 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.094299 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.094338 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.094417 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.100073 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.123021 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.148068 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.181036 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.196921 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.196990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.197010 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.197034 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.197051 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.209014 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.237337 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.258942 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.275940 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.288042 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.299263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.299294 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.299304 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.299323 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.299333 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.300760 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.314346 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:06Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.401484 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.401516 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.401526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.401540 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.401549 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.504704 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.504800 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.504816 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.504839 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.504858 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.608252 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.608317 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.608332 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.608355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.608370 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.711836 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.711895 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.711913 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.711940 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.711957 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.736065 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.736174 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:06 crc kubenswrapper[4817]: E1128 14:30:06.736392 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:06 crc kubenswrapper[4817]: E1128 14:30:06.736191 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.814319 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.814354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.814364 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.814380 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.814392 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.917907 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.917993 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.918014 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.918041 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.918060 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:06Z","lastTransitionTime":"2025-11-28T14:30:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:06 crc kubenswrapper[4817]: I1128 14:30:06.994746 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.022273 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.022335 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.022353 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.022379 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.022399 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.126524 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.126595 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.126615 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.126644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.126661 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.230131 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.230205 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.230217 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.230236 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.230249 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.333302 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.333383 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.333392 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.333406 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.333421 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.436437 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.436555 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.436574 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.436599 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.436616 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.539753 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.539826 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.539850 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.539882 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.539899 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.643372 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.643432 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.643445 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.643469 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.643483 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.736411 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:07 crc kubenswrapper[4817]: E1128 14:30:07.736610 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.746129 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.746205 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.746221 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.746239 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.746250 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.759826 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.781273 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.812260 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.840465 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.848569 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.848616 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.848629 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.848648 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.848660 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.853210 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.867182 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.882951 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.897548 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.911149 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.925389 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.940682 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.951610 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.951688 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.951714 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.951792 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.951819 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:07Z","lastTransitionTime":"2025-11-28T14:30:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.957011 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.969290 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.992068 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:07 crc kubenswrapper[4817]: I1128 14:30:07.999697 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/0.log" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.002301 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee" exitCode=1 Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.002347 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.003158 4817 scope.go:117] "RemoveContainer" containerID="e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.010518 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.028566 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.040903 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.054284 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.054484 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.054612 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.054688 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.054791 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.064122 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.079280 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.093104 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.103204 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.114276 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.129214 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.144045 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.157549 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.157599 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.157616 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.157640 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.157656 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.167951 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:07Z\\\",\\\"message\\\":\\\"30:07.273178 6146 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 14:30:07.273188 6146 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 14:30:07.273248 6146 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 14:30:07.273292 6146 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273425 6146 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273501 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273715 6146 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.274454 6146 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 14:30:07.274689 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.275070 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 14:30:07.275122 6146 factory.go:656] Stopping watch factory\\\\nI1128 14:30:07.275142 6146 ovnkube.go:599] Stopped ovnkube\\\\nI1128 14:30:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.185245 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.197642 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.213598 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.226194 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.238395 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.259991 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.260036 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.260055 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.260078 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.260095 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.363173 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.363212 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.363221 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.363238 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.363247 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.465716 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.465776 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.465791 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.465811 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.465821 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.568488 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.569074 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.569085 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.569105 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.569118 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.671464 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.671511 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.671528 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.671551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.671570 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.736391 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.736391 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:08 crc kubenswrapper[4817]: E1128 14:30:08.736527 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:08 crc kubenswrapper[4817]: E1128 14:30:08.736802 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.773949 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.774002 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.774022 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.774047 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.774064 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.877075 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.877113 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.877125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.877143 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.877154 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.980698 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.980784 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.980804 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.980828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:08 crc kubenswrapper[4817]: I1128 14:30:08.980846 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:08Z","lastTransitionTime":"2025-11-28T14:30:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.014065 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/0.log" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.018165 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.018355 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.043564 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.067771 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.084046 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.084107 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.084125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.084151 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.084170 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.099222 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.120616 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.139851 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.173378 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.186603 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.186671 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.186688 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.186713 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.186760 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.194977 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.216352 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.236292 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.268681 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:07Z\\\",\\\"message\\\":\\\"30:07.273178 6146 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 14:30:07.273188 6146 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 14:30:07.273248 6146 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 14:30:07.273292 6146 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273425 6146 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273501 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273715 6146 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.274454 6146 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 14:30:07.274689 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.275070 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 14:30:07.275122 6146 factory.go:656] Stopping watch factory\\\\nI1128 14:30:07.275142 6146 ovnkube.go:599] Stopped ovnkube\\\\nI1128 14:30:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.270369 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.270437 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.270458 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.270486 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.270504 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.289433 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.295526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.295591 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.295611 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.295639 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.295659 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.296213 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.308920 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.323402 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.328705 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.330414 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.330446 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.330457 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.330476 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.330487 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.344300 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.346417 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.349592 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.349637 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.349654 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.349679 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.349697 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.364738 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.365809 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.368702 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.368756 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.368767 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.368784 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.368794 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.381346 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:09Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.381508 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.382996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.383029 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.383040 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.383058 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.383071 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.485120 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.485171 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.485182 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.485200 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.485210 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.588045 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.588104 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.588128 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.588235 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.588266 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.691950 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.692003 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.692026 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.692050 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.692065 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.737085 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:09 crc kubenswrapper[4817]: E1128 14:30:09.737257 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.795378 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.795475 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.795491 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.795514 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.795530 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.898570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.898639 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.898657 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.898682 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:09 crc kubenswrapper[4817]: I1128 14:30:09.898703 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:09Z","lastTransitionTime":"2025-11-28T14:30:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.001843 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.001907 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.001923 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.001945 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.001959 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.023606 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/1.log" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.024345 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/0.log" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.027975 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac" exitCode=1 Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.028052 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.028139 4817 scope.go:117] "RemoveContainer" containerID="e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.029439 4817 scope.go:117] "RemoveContainer" containerID="a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac" Nov 28 14:30:10 crc kubenswrapper[4817]: E1128 14:30:10.029807 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.051336 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.069685 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.084244 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.097368 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.104133 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.104168 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.104180 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.104198 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.104211 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.113976 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.137613 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.153578 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.175460 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.190334 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.207216 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.207266 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.207289 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.207320 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.207340 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.209601 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.235534 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.253033 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.281644 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:07Z\\\",\\\"message\\\":\\\"30:07.273178 6146 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 14:30:07.273188 6146 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 14:30:07.273248 6146 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 14:30:07.273292 6146 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273425 6146 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273501 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273715 6146 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.274454 6146 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 14:30:07.274689 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.275070 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 14:30:07.275122 6146 factory.go:656] Stopping watch factory\\\\nI1128 14:30:07.275142 6146 ovnkube.go:599] Stopped ovnkube\\\\nI1128 14:30:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.299071 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.310408 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.310450 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.310462 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.310481 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.310493 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.311591 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.414122 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.414200 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.414222 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.414249 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.414267 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.517384 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.517477 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.517502 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.517533 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.517568 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.620607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.620652 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.620665 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.620743 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.620753 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.724161 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.724214 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.724232 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.724259 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.724280 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.736910 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:10 crc kubenswrapper[4817]: E1128 14:30:10.737108 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.737384 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:10 crc kubenswrapper[4817]: E1128 14:30:10.737610 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.801192 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8"] Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.801931 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.804152 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.804891 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.820132 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgv7l\" (UniqueName: \"kubernetes.io/projected/7d3ad39a-79e7-4547-a004-00e2826f1bd2-kube-api-access-xgv7l\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.820221 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d3ad39a-79e7-4547-a004-00e2826f1bd2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.820444 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d3ad39a-79e7-4547-a004-00e2826f1bd2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.820549 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d3ad39a-79e7-4547-a004-00e2826f1bd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.827015 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.827069 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.827081 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.827107 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.827121 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.841007 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.868751 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.890941 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.905180 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.919565 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.921990 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgv7l\" (UniqueName: \"kubernetes.io/projected/7d3ad39a-79e7-4547-a004-00e2826f1bd2-kube-api-access-xgv7l\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.922044 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d3ad39a-79e7-4547-a004-00e2826f1bd2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.922110 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d3ad39a-79e7-4547-a004-00e2826f1bd2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.922146 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d3ad39a-79e7-4547-a004-00e2826f1bd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.923135 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d3ad39a-79e7-4547-a004-00e2826f1bd2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.923292 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d3ad39a-79e7-4547-a004-00e2826f1bd2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.928901 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d3ad39a-79e7-4547-a004-00e2826f1bd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.930681 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.930895 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.930952 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.930964 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.930984 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.930997 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:10Z","lastTransitionTime":"2025-11-28T14:30:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.938686 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgv7l\" (UniqueName: \"kubernetes.io/projected/7d3ad39a-79e7-4547-a004-00e2826f1bd2-kube-api-access-xgv7l\") pod \"ovnkube-control-plane-749d76644c-pv4m8\" (UID: \"7d3ad39a-79e7-4547-a004-00e2826f1bd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.944780 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.958541 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.983642 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:10 crc kubenswrapper[4817]: I1128 14:30:10.999708 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:10Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.015867 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.034338 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.034404 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.034416 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.034454 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.034465 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.036590 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/1.log" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.040819 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.067606 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:07Z\\\",\\\"message\\\":\\\"30:07.273178 6146 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 14:30:07.273188 6146 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 14:30:07.273248 6146 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 14:30:07.273292 6146 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273425 6146 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273501 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273715 6146 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.274454 6146 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 14:30:07.274689 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.275070 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 14:30:07.275122 6146 factory.go:656] Stopping watch factory\\\\nI1128 14:30:07.275142 6146 ovnkube.go:599] Stopped ovnkube\\\\nI1128 14:30:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.085519 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.099078 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.115030 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.117202 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" Nov 28 14:30:11 crc kubenswrapper[4817]: W1128 14:30:11.131233 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d3ad39a_79e7_4547_a004_00e2826f1bd2.slice/crio-0eea25126fd68d0da42f0d438e3f2fc11c8e9e4832dc5dd93057979432d1a06d WatchSource:0}: Error finding container 0eea25126fd68d0da42f0d438e3f2fc11c8e9e4832dc5dd93057979432d1a06d: Status 404 returned error can't find the container with id 0eea25126fd68d0da42f0d438e3f2fc11c8e9e4832dc5dd93057979432d1a06d Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.136376 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.136422 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.136440 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.136469 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.136489 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.240158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.240236 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.240249 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.240267 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.240278 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.343336 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.343383 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.343392 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.343413 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.343423 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.446687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.446811 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.446830 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.446855 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.446872 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.542314 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zvnxb"] Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.543039 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:11 crc kubenswrapper[4817]: E1128 14:30:11.543137 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.549636 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.549675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.549687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.549706 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.549752 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.556017 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.576463 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.597254 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.612791 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.628009 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.630424 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.630473 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggdn\" (UniqueName: \"kubernetes.io/projected/bfd936ce-70b3-4e00-944f-f505238d1ab2-kube-api-access-2ggdn\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.641001 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.653826 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.653861 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.653871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.653886 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.653897 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.658989 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.670221 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.682700 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.696682 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.709186 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.732039 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggdn\" (UniqueName: \"kubernetes.io/projected/bfd936ce-70b3-4e00-944f-f505238d1ab2-kube-api-access-2ggdn\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.732194 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:11 crc kubenswrapper[4817]: E1128 14:30:11.732353 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:11 crc kubenswrapper[4817]: E1128 14:30:11.732435 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:12.232410408 +0000 UTC m=+34.820388684 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.736500 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:11 crc kubenswrapper[4817]: E1128 14:30:11.736639 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.745534 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:07Z\\\",\\\"message\\\":\\\"30:07.273178 6146 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 14:30:07.273188 6146 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 14:30:07.273248 6146 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 14:30:07.273292 6146 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273425 6146 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273501 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273715 6146 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.274454 6146 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 14:30:07.274689 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.275070 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 14:30:07.275122 6146 factory.go:656] Stopping watch factory\\\\nI1128 14:30:07.275142 6146 ovnkube.go:599] Stopped ovnkube\\\\nI1128 14:30:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.753672 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggdn\" (UniqueName: \"kubernetes.io/projected/bfd936ce-70b3-4e00-944f-f505238d1ab2-kube-api-access-2ggdn\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.755868 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.755892 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.755900 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.755936 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.755946 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.761505 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.776646 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.791302 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.805304 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.821795 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:11Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.858802 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.858845 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.858855 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.858872 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.858883 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.961895 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.961963 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.961982 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.962011 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:11 crc kubenswrapper[4817]: I1128 14:30:11.962030 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:11Z","lastTransitionTime":"2025-11-28T14:30:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.049023 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" event={"ID":"7d3ad39a-79e7-4547-a004-00e2826f1bd2","Type":"ContainerStarted","Data":"80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.049120 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" event={"ID":"7d3ad39a-79e7-4547-a004-00e2826f1bd2","Type":"ContainerStarted","Data":"597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.049159 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" event={"ID":"7d3ad39a-79e7-4547-a004-00e2826f1bd2","Type":"ContainerStarted","Data":"0eea25126fd68d0da42f0d438e3f2fc11c8e9e4832dc5dd93057979432d1a06d"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.065164 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.065239 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.065258 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.065287 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.065310 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.068038 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.101866 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.121035 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.134811 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.135878 4817 scope.go:117] "RemoveContainer" containerID="a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac" Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.136109 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.138970 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.157151 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.167508 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.167586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.167613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.167648 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.167666 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.178011 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.198887 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.213629 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.231707 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.239108 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.239307 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.239422 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:13.23938733 +0000 UTC m=+35.827365716 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.252443 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.270430 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.270469 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.270484 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.270511 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.270526 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.271301 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.301800 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3e88e9e306394c617edeb69c0fa6d1651032d9c5d61c60913116cee2d7c25ee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:07Z\\\",\\\"message\\\":\\\"30:07.273178 6146 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 14:30:07.273188 6146 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 14:30:07.273248 6146 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 14:30:07.273292 6146 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273425 6146 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273501 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.273715 6146 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.274454 6146 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 14:30:07.274689 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1128 14:30:07.275070 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 14:30:07.275122 6146 factory.go:656] Stopping watch factory\\\\nI1128 14:30:07.275142 6146 ovnkube.go:599] Stopped ovnkube\\\\nI1128 14:30:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.323955 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.340618 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.362020 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.373947 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.374042 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.374068 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.374099 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.374121 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.379686 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.401957 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.434017 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.440675 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.440988 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:30:28.440921279 +0000 UTC m=+51.028899565 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.441094 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.441303 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.441371 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.441446 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.441822 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.441927 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.441954 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.442015 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.442047 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.441992 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.442134 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.442172 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.442923 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:28.442868652 +0000 UTC m=+51.030846988 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.442988 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:28.442961064 +0000 UTC m=+51.030939370 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.443086 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:28.443015256 +0000 UTC m=+51.030993562 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.443177 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:28.443131539 +0000 UTC m=+51.031109845 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.456002 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.470944 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.476584 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.476632 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.476645 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.476664 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.476678 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.490292 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.503387 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.516193 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.533348 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.547664 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.561173 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.579202 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.579240 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.579251 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.579269 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.579279 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.579338 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.599145 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.612636 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.627981 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.656588 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.679908 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.683234 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.683289 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.683310 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.683338 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.683356 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.698174 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.711817 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:12Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.736312 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.736368 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.737082 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:12 crc kubenswrapper[4817]: E1128 14:30:12.737240 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.786929 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.786973 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.786985 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.787005 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.787019 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.889662 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.889737 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.889755 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.889778 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.889794 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.993469 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.993547 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.993566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.993595 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:12 crc kubenswrapper[4817]: I1128 14:30:12.993616 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:12Z","lastTransitionTime":"2025-11-28T14:30:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.096426 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.096496 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.096521 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.096551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.096573 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.199875 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.199941 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.199962 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.199990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.200014 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.252276 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:13 crc kubenswrapper[4817]: E1128 14:30:13.252495 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:13 crc kubenswrapper[4817]: E1128 14:30:13.252605 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:15.252572456 +0000 UTC m=+37.840550762 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.303618 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.303707 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.303775 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.303802 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.303815 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.406643 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.406687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.406704 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.406767 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.406807 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.509992 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.510070 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.510093 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.510128 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.510152 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.613135 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.613197 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.613216 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.613245 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.613264 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.715544 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.715631 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.715649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.715675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.715693 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.736250 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.736347 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:13 crc kubenswrapper[4817]: E1128 14:30:13.736491 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:13 crc kubenswrapper[4817]: E1128 14:30:13.736648 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.818799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.818872 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.818889 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.818914 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.818933 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.928144 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.928221 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.928238 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.928265 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:13 crc kubenswrapper[4817]: I1128 14:30:13.928286 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:13Z","lastTransitionTime":"2025-11-28T14:30:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.031962 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.032021 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.032041 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.032068 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.032085 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.135379 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.135448 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.135468 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.135500 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.135519 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.239242 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.239297 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.239318 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.239344 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.239363 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.344516 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.344621 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.344677 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.344750 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.344769 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.448623 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.448685 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.448701 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.448753 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.448771 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.552164 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.552215 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.552233 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.552256 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.552274 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.655087 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.655179 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.655207 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.655240 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.655265 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.736693 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.736774 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:14 crc kubenswrapper[4817]: E1128 14:30:14.736874 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:14 crc kubenswrapper[4817]: E1128 14:30:14.736998 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.757953 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.757990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.758000 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.758018 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.758029 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.861655 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.861762 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.861788 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.861816 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.861837 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.965038 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.965111 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.965137 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.965166 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:14 crc kubenswrapper[4817]: I1128 14:30:14.965190 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:14Z","lastTransitionTime":"2025-11-28T14:30:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.068480 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.068556 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.068577 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.068621 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.068640 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.171175 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.171233 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.171256 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.171286 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.171307 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.274024 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.274096 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.274130 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.274160 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.274183 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.276903 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:15 crc kubenswrapper[4817]: E1128 14:30:15.277120 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:15 crc kubenswrapper[4817]: E1128 14:30:15.277240 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:19.277207561 +0000 UTC m=+41.865185867 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.378096 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.378176 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.378200 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.378226 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.378243 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.481226 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.481303 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.481329 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.481359 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.481397 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.584459 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.584501 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.584520 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.584548 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.584569 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.687788 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.687832 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.687850 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.687882 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.687896 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.736589 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.736587 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:15 crc kubenswrapper[4817]: E1128 14:30:15.736755 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:15 crc kubenswrapper[4817]: E1128 14:30:15.736949 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.791148 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.791562 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.791705 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.791886 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.792022 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.895692 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.895811 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.895846 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.895877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.895899 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.999333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.999398 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.999420 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.999455 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:15 crc kubenswrapper[4817]: I1128 14:30:15.999481 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:15Z","lastTransitionTime":"2025-11-28T14:30:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.102104 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.102185 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.102208 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.102240 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.102261 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.205221 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.205286 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.205303 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.205327 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.205345 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.308371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.308455 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.308477 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.308507 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.308527 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.412669 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.412756 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.412767 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.412785 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.412799 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.516224 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.516328 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.516345 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.516372 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.516391 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.521002 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.543659 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.577177 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.602286 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.620002 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.620053 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.620071 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.620100 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.620121 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.622394 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.640284 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.660881 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.684883 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.705080 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.723986 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.724028 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.724037 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.724056 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.724069 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.728393 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.736449 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.736493 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:16 crc kubenswrapper[4817]: E1128 14:30:16.736642 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:16 crc kubenswrapper[4817]: E1128 14:30:16.736837 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.746032 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.761953 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.777279 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.793609 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.806696 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.826700 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.826768 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.826778 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.826794 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.826803 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.826965 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.846756 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.864508 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:16Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.929969 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.930033 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.930051 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.930075 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:16 crc kubenswrapper[4817]: I1128 14:30:16.930094 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:16Z","lastTransitionTime":"2025-11-28T14:30:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.032468 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.032542 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.032555 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.032586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.032601 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.136309 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.136395 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.136407 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.136436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.136448 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.239019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.239081 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.239098 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.239125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.239143 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.342384 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.342449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.342467 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.342495 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.342511 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.445570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.445672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.445696 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.445758 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.445786 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.548705 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.548793 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.548832 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.548871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.548902 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.652329 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.652399 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.652425 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.652457 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.652487 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.736468 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.736480 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:17 crc kubenswrapper[4817]: E1128 14:30:17.736713 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:17 crc kubenswrapper[4817]: E1128 14:30:17.736808 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.755372 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.755452 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.755478 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.755512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.755539 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.759936 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.779198 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.801600 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.827637 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.850913 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.859009 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.859053 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.859065 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.859086 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.859101 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.865886 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.881494 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.902999 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.917084 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.936852 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.947381 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.962130 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.962177 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.962190 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.962211 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.962231 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:17Z","lastTransitionTime":"2025-11-28T14:30:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.969738 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:17 crc kubenswrapper[4817]: I1128 14:30:17.988838 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:17Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.002553 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:18Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.013907 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:18Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.031502 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:18Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.049111 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:18Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.064949 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.064999 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.065021 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.065048 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.065065 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.168101 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.168163 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.168180 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.168204 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.168222 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.271200 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.271271 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.271280 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.271305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.271317 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.374443 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.374514 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.374533 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.374559 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.374577 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.478488 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.478549 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.478568 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.478592 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.478611 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.581077 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.581143 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.581162 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.581188 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.581206 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.683877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.683927 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.683945 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.683969 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.683987 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.736664 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:18 crc kubenswrapper[4817]: E1128 14:30:18.736920 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.736673 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:18 crc kubenswrapper[4817]: E1128 14:30:18.737228 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.786996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.787071 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.787096 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.787121 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.787139 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.891040 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.891113 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.891128 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.891152 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.891168 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.994512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.994589 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.994613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.994645 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:18 crc kubenswrapper[4817]: I1128 14:30:18.994674 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:18Z","lastTransitionTime":"2025-11-28T14:30:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.098260 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.098323 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.098343 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.098367 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.098385 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.201204 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.201263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.201279 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.201304 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.201320 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.304286 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.304363 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.304387 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.304416 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.304437 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.328188 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.328371 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.328509 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:27.328440345 +0000 UTC m=+49.916418641 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.408056 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.408124 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.408141 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.408172 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.408190 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.511566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.511633 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.511650 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.511673 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.511690 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.539832 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.539897 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.539915 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.539944 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.539964 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.561897 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:19Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.567297 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.567346 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.567364 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.567388 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.567404 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.589456 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:19Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.595349 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.595417 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.595440 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.595467 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.595485 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.615269 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:19Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.620835 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.620902 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.620925 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.620954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.620977 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.638799 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:19Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.644090 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.644129 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.644145 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.644166 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.644181 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.660414 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:19Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.660562 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.662360 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.662392 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.662405 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.662421 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.662433 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.736919 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.736996 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.737060 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:19 crc kubenswrapper[4817]: E1128 14:30:19.737185 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.767743 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.767786 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.767797 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.767812 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.767821 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.870384 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.870422 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.870435 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.870449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.870460 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.973182 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.973252 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.973271 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.973295 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:19 crc kubenswrapper[4817]: I1128 14:30:19.973313 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:19Z","lastTransitionTime":"2025-11-28T14:30:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.076634 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.076709 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.076758 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.076790 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.076827 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.180005 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.180088 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.180111 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.180148 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.180172 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.283485 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.283538 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.283552 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.283572 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.283585 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.385958 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.386002 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.386012 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.386025 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.386035 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.489908 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.489972 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.489989 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.490014 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.490031 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.592930 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.592996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.593015 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.593040 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.593057 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.696105 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.696175 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.696198 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.696228 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.696252 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.735924 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.736024 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:20 crc kubenswrapper[4817]: E1128 14:30:20.736108 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:20 crc kubenswrapper[4817]: E1128 14:30:20.736209 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.799483 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.799550 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.799569 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.799596 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.799619 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.902295 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.902362 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.902388 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.902423 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:20 crc kubenswrapper[4817]: I1128 14:30:20.902449 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:20Z","lastTransitionTime":"2025-11-28T14:30:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.005760 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.005840 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.005860 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.005885 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.005906 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.108656 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.108716 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.108762 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.108786 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.108803 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.211797 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.211893 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.211915 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.212257 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.212529 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.315931 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.315994 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.316017 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.316045 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.316065 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.419545 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.419586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.419596 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.419613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.419626 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.522424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.522501 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.522524 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.522553 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.522573 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.624811 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.624854 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.624866 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.624884 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.624895 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.727528 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.727588 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.727609 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.727628 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.727642 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.736033 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.736040 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:21 crc kubenswrapper[4817]: E1128 14:30:21.736237 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:21 crc kubenswrapper[4817]: E1128 14:30:21.736307 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.830838 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.830907 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.830929 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.830959 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.830979 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.934065 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.934129 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.934149 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.934175 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:21 crc kubenswrapper[4817]: I1128 14:30:21.934198 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:21Z","lastTransitionTime":"2025-11-28T14:30:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.037146 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.037551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.037689 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.037872 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.038099 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.140862 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.140930 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.141512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.143380 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.143414 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.247352 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.247422 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.247443 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.247473 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.247495 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.351219 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.351271 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.351284 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.351308 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.351323 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.453987 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.454074 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.454087 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.454134 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.454149 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.556837 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.556889 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.556901 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.556920 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.556942 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.660848 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.660902 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.660917 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.660941 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.660955 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.736462 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.736492 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:22 crc kubenswrapper[4817]: E1128 14:30:22.736696 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:22 crc kubenswrapper[4817]: E1128 14:30:22.736816 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.764283 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.764354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.764369 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.764393 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.764405 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.867011 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.867082 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.867097 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.867121 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.867137 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.970056 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.970123 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.970140 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.970168 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:22 crc kubenswrapper[4817]: I1128 14:30:22.970183 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:22Z","lastTransitionTime":"2025-11-28T14:30:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.073955 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.074007 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.074020 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.074042 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.074056 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.176235 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.176292 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.176308 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.176334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.176351 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.206287 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.217672 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.223398 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.240201 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.263931 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.279791 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.279867 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.279881 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.279904 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.279919 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.282510 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.298887 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.315412 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.334908 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.348937 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.365876 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.382837 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.383228 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.383260 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.383271 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.383288 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.383299 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.400217 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.416989 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.443504 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.460364 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.482383 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.487284 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.487355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.487373 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.487401 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.487423 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.499017 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.519552 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:23Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.592270 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.592355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.592376 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.592404 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.592425 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.696516 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.696929 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.696956 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.696980 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.696998 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.736315 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.736372 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:23 crc kubenswrapper[4817]: E1128 14:30:23.736482 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:23 crc kubenswrapper[4817]: E1128 14:30:23.736881 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.801305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.801366 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.801377 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.801399 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.801416 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.905081 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.905437 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.905585 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.905794 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:23 crc kubenswrapper[4817]: I1128 14:30:23.906000 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:23Z","lastTransitionTime":"2025-11-28T14:30:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.009851 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.009928 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.009954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.009983 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.010006 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.113398 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.113484 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.113512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.113541 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.113564 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.217103 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.217178 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.217192 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.217223 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.217243 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.320274 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.320328 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.320339 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.320361 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.320375 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.424754 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.424847 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.424861 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.424888 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.424905 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.527793 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.527849 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.527861 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.527883 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.527897 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.631847 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.631912 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.631934 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.631964 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.631984 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735616 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735680 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735698 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735754 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735772 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735908 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.735986 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:24 crc kubenswrapper[4817]: E1128 14:30:24.736076 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:24 crc kubenswrapper[4817]: E1128 14:30:24.736206 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.839342 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.839433 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.839444 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.839465 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.839478 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.942209 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.942257 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.942266 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.942291 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:24 crc kubenswrapper[4817]: I1128 14:30:24.942314 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:24Z","lastTransitionTime":"2025-11-28T14:30:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.045170 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.045248 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.045274 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.045305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.045330 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.149020 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.149089 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.149118 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.149159 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.149194 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.253241 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.253326 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.253340 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.253357 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.253369 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.356700 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.356786 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.356807 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.356831 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.356849 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.459841 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.459921 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.459947 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.459978 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.460003 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.563019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.563821 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.563842 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.563859 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.563875 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.667335 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.667413 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.667436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.667465 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.667488 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.736987 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.737033 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:25 crc kubenswrapper[4817]: E1128 14:30:25.737222 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:25 crc kubenswrapper[4817]: E1128 14:30:25.737401 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.770341 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.770401 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.770419 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.770444 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.770465 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.874091 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.874158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.874177 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.874217 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.874237 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.977053 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.977155 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.977176 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.977200 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:25 crc kubenswrapper[4817]: I1128 14:30:25.977219 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:25Z","lastTransitionTime":"2025-11-28T14:30:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.080827 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.080885 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.080904 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.080927 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.080944 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.184145 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.184221 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.184239 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.184263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.184281 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.287955 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.288022 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.288039 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.288069 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.288087 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.391625 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.391687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.391699 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.391748 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.391764 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.495356 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.495410 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.495419 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.495438 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.495455 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.598415 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.598694 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.598762 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.598798 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.598820 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.702863 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.702940 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.702961 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.702991 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.703014 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.736986 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.737116 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:26 crc kubenswrapper[4817]: E1128 14:30:26.737322 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:26 crc kubenswrapper[4817]: E1128 14:30:26.737427 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.738534 4817 scope.go:117] "RemoveContainer" containerID="a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.806263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.806332 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.806354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.806380 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.806399 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.909134 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.909669 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.909682 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.909703 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:26 crc kubenswrapper[4817]: I1128 14:30:26.909746 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:26Z","lastTransitionTime":"2025-11-28T14:30:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.012669 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.012728 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.012741 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.012760 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.012778 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.107981 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/1.log" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.110939 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.111895 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.115583 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.115646 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.115665 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.115693 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.115743 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.134616 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.153044 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.173361 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.192295 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.215525 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.218530 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.218576 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.218587 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.218607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.218616 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.232529 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.250638 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.274991 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.292238 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.313204 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.320812 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.320857 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.320868 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.320883 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.320892 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.330448 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.342710 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.353870 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.368114 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.381756 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.397545 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.413224 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.422347 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:27 crc kubenswrapper[4817]: E1128 14:30:27.422511 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:27 crc kubenswrapper[4817]: E1128 14:30:27.422590 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:30:43.422568807 +0000 UTC m=+66.010547273 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.423457 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.423503 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.423515 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.423529 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.423539 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.434161 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.525566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.525616 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.525630 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.525646 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.525658 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.627445 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.627484 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.627492 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.627506 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.627516 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.730502 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.730589 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.730613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.730649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.730677 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.736877 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.737029 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:27 crc kubenswrapper[4817]: E1128 14:30:27.737139 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:27 crc kubenswrapper[4817]: E1128 14:30:27.737259 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.754249 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.766239 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.782356 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.795894 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.811111 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.822226 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.832472 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.832519 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.832531 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.832546 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.832557 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.839369 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.853307 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.871663 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.882454 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.893379 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.904387 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.921774 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.935199 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.935247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.935263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.935284 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.935295 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:27Z","lastTransitionTime":"2025-11-28T14:30:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.942431 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.954758 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.966655 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.982200 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:27 crc kubenswrapper[4817]: I1128 14:30:27.998999 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:27Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.039247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.039321 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.039340 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.039367 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.039384 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.117050 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/2.log" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.118364 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/1.log" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.121580 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e" exitCode=1 Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.121627 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.121676 4817 scope.go:117] "RemoveContainer" containerID="a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.122645 4817 scope.go:117] "RemoveContainer" containerID="53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e" Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.122889 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.138797 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.142475 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.142542 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.142559 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.142585 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.142604 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.154403 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.167314 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.177912 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.188799 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.206979 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.222563 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.242841 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.245112 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.245211 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.245232 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.245262 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.245286 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.258289 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.278294 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8b2cc1da9e55b01d5f4d12391071a8d9c18d07b6357c4c58e42763187e831ac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:09Z\\\",\\\"message\\\":\\\"bj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1128 14:30:08.801378 6275 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1128 14:30:08.801409 6275 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}\\\\nF1128 14:30:08.801430 6275 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:08Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:08.801436 6275 obj_retry.go:303] Retry object se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.298386 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.313918 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.327891 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.342660 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.347541 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.347576 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.347588 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.347606 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.347617 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.364434 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.387112 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.405005 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.420555 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:28Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.451032 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.451082 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.451092 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.451110 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.451121 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.537371 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.537504 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.537553 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.537683 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.537692 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:31:00.537616352 +0000 UTC m=+83.125594648 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.537707 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.537694 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.537774 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.538029 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:31:00.537994892 +0000 UTC m=+83.125973198 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.537734 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.538077 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.538200 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.538332 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.538357 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.538103 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.538421 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:31:00.538345692 +0000 UTC m=+83.126323958 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.539227 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:31:00.539203355 +0000 UTC m=+83.127181621 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.539265 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:31:00.539256226 +0000 UTC m=+83.127234492 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.554985 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.555061 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.555085 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.555115 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.555143 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.658111 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.658177 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.658199 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.658225 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.658243 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.736542 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.736629 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.736773 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:28 crc kubenswrapper[4817]: E1128 14:30:28.736938 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.760580 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.760660 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.760686 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.760747 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.760772 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.863803 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.863859 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.863870 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.863889 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.863909 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.966644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.966695 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.966709 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.966750 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:28 crc kubenswrapper[4817]: I1128 14:30:28.966762 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:28Z","lastTransitionTime":"2025-11-28T14:30:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.069367 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.069419 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.069431 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.069456 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.069468 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.126649 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/2.log" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.130858 4817 scope.go:117] "RemoveContainer" containerID="53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e" Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.131227 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.145672 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.168060 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.171933 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.171970 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.171996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.172016 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.172029 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.183091 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.197196 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.213034 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.231340 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.246434 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.262299 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.274935 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.274997 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.275015 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.275041 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.275059 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.275254 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.289865 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.323399 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.338357 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.354703 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.375659 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.377637 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.377678 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.377691 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.377708 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.377782 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.392987 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.404333 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.416595 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.431646 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.480344 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.480397 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.480415 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.480441 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.480463 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.584970 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.585035 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.585052 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.585076 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.585093 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.687884 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.687961 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.687981 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.688005 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.688022 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.736668 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.736811 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.736940 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.737062 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.791587 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.791672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.791698 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.791777 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.791821 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.795987 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.796045 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.796063 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.796087 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.796104 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.819846 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.825407 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.825464 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.825483 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.825505 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.825522 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.845150 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.850348 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.850392 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.850403 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.850421 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.850435 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.866834 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.871220 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.871264 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.871275 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.871292 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.871306 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.889364 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.894356 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.894406 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.894424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.894451 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.894472 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.913547 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:29Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:29 crc kubenswrapper[4817]: E1128 14:30:29.913815 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.915933 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.915974 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.915985 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.916003 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:29 crc kubenswrapper[4817]: I1128 14:30:29.916014 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:29Z","lastTransitionTime":"2025-11-28T14:30:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.019334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.019455 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.019481 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.019526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.019553 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.122275 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.122440 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.122467 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.122507 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.122581 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.225311 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.225376 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.225405 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.225436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.225459 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.328599 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.328636 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.328647 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.328663 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.328673 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.431438 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.431503 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.431521 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.431546 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.431565 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.534903 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.534968 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.534985 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.535010 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.535027 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.638167 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.638222 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.638241 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.638263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.638280 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.736956 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.736965 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:30 crc kubenswrapper[4817]: E1128 14:30:30.737153 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:30 crc kubenswrapper[4817]: E1128 14:30:30.737324 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.741695 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.741783 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.741807 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.741833 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.741860 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.845023 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.845087 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.845105 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.845129 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.845145 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.948561 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.948635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.948653 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.948684 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:30 crc kubenswrapper[4817]: I1128 14:30:30.948701 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:30Z","lastTransitionTime":"2025-11-28T14:30:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.051165 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.051227 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.051249 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.051281 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.051305 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.154494 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.154570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.154654 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.154755 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.154786 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.258415 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.258484 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.258494 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.258509 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.258519 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.361994 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.362054 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.362071 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.362098 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.362117 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.465270 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.465335 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.465352 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.465379 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.465396 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.568042 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.568094 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.568111 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.568135 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.568153 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.671235 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.671287 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.671303 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.671325 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.671342 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.736101 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.736243 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:31 crc kubenswrapper[4817]: E1128 14:30:31.736296 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:31 crc kubenswrapper[4817]: E1128 14:30:31.736449 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.776276 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.776343 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.776366 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.776391 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.776407 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.878921 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.878964 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.878978 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.878993 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.879002 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.980702 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.980826 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.980847 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.980876 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:31 crc kubenswrapper[4817]: I1128 14:30:31.980899 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:31Z","lastTransitionTime":"2025-11-28T14:30:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.083884 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.083948 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.083968 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.083992 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.084009 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.187852 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.187925 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.187951 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.187983 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.188006 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.290994 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.291038 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.291053 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.291073 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.291089 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.396329 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.396444 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.396461 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.396498 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.396519 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.499426 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.499489 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.499506 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.499529 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.499546 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.603041 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.603118 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.603131 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.603176 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.603193 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.706512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.706580 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.706602 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.706634 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.706657 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.736525 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.736633 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:32 crc kubenswrapper[4817]: E1128 14:30:32.736705 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:32 crc kubenswrapper[4817]: E1128 14:30:32.736877 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.809995 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.810056 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.810075 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.810099 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.810115 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.912574 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.912633 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.912650 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.912675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:32 crc kubenswrapper[4817]: I1128 14:30:32.912692 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:32Z","lastTransitionTime":"2025-11-28T14:30:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.015707 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.015811 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.015835 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.015863 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.015882 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.119275 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.119332 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.119348 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.119373 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.119392 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.222051 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.222109 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.222126 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.222149 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.222166 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.325858 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.325950 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.325986 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.326019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.326042 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.428809 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.428935 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.428956 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.428980 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.428997 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.532376 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.532435 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.532453 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.532481 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.532498 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.635774 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.635841 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.635863 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.635892 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.635915 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.736958 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.736958 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:33 crc kubenswrapper[4817]: E1128 14:30:33.737272 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:33 crc kubenswrapper[4817]: E1128 14:30:33.737241 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.738658 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.738677 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.738685 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.738695 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.738704 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.841427 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.841489 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.841510 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.841533 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.841552 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.948401 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.948996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.949202 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.949229 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:33 crc kubenswrapper[4817]: I1128 14:30:33.949248 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:33Z","lastTransitionTime":"2025-11-28T14:30:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.052989 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.053261 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.053285 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.053315 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.053339 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.156832 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.156911 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.156936 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.156967 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.156990 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.261228 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.261288 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.261305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.261333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.261357 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.365123 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.365437 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.365456 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.365480 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.365530 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.469253 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.469328 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.469345 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.469368 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.469410 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.572375 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.572447 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.572471 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.572532 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.572559 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.675827 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.675886 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.675906 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.675931 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.675948 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.736874 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.736941 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:34 crc kubenswrapper[4817]: E1128 14:30:34.737070 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:34 crc kubenswrapper[4817]: E1128 14:30:34.737197 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.779365 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.779425 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.779449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.779477 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.779500 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.882423 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.882488 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.882508 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.882535 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.882552 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.985438 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.985492 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.985510 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.985532 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:34 crc kubenswrapper[4817]: I1128 14:30:34.985549 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:34Z","lastTransitionTime":"2025-11-28T14:30:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.088926 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.088970 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.088988 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.089010 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.089026 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.191672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.191759 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.191778 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.191803 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.191820 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.294697 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.294808 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.294849 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.294878 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.294900 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.397223 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.397297 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.397319 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.397350 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.397375 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.500526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.500586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.500608 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.500639 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.500659 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.603496 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.603564 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.603585 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.603611 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.603627 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.706447 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.706507 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.706525 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.706551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.706568 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.736644 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.737219 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:35 crc kubenswrapper[4817]: E1128 14:30:35.737386 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:35 crc kubenswrapper[4817]: E1128 14:30:35.737569 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.809687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.809801 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.809828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.809860 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.809883 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.912833 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.912916 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.912937 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.912962 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:35 crc kubenswrapper[4817]: I1128 14:30:35.912981 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:35Z","lastTransitionTime":"2025-11-28T14:30:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.016236 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.016317 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.016341 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.016373 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.016410 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.119516 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.119617 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.119635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.119659 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.119677 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.223230 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.223354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.223380 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.223409 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.223433 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.326743 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.326799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.326819 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.326848 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.326872 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.429566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.429638 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.429667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.429711 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.429783 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.532707 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.532793 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.532810 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.532835 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.532852 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.635802 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.636243 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.636408 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.636570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.636704 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.735988 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.736030 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:36 crc kubenswrapper[4817]: E1128 14:30:36.736173 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:36 crc kubenswrapper[4817]: E1128 14:30:36.736339 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.740767 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.740835 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.740853 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.740877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.740895 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.844284 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.844347 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.844372 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.844399 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.844420 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.948887 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.948954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.948973 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.948998 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:36 crc kubenswrapper[4817]: I1128 14:30:36.949018 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:36Z","lastTransitionTime":"2025-11-28T14:30:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.052120 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.052208 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.052233 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.052264 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.052283 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.155457 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.155532 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.155551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.155579 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.155598 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.258503 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.258572 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.258591 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.258628 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.258646 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.362134 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.362236 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.362255 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.362281 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.362300 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.464651 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.464687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.464696 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.464709 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.464742 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.567258 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.567871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.567900 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.567913 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.567923 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.671309 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.671392 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.671416 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.671449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.671472 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.736007 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.736166 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:37 crc kubenswrapper[4817]: E1128 14:30:37.736348 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:37 crc kubenswrapper[4817]: E1128 14:30:37.736514 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.764945 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.774930 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.775002 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.775023 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.775051 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.775069 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.787906 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.806156 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.829926 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.855005 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.874278 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.877617 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.877920 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.878088 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.878301 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.878467 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.902330 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.918769 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.933127 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.944768 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.959106 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.980424 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.981893 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.981964 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.981993 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.982021 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.982044 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:37Z","lastTransitionTime":"2025-11-28T14:30:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:37 crc kubenswrapper[4817]: I1128 14:30:37.993294 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:37Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.005549 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:38Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.019022 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:38Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.030355 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:38Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.043229 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:38Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.063242 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:38Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.084814 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.084872 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.084890 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.084913 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.084930 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.188317 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.188381 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.188398 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.188424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.188442 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.291178 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.291249 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.291267 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.291293 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.291312 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.394777 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.394854 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.394881 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.394911 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.394932 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.497918 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.497999 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.498021 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.498048 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.498065 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.600579 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.600645 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.600661 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.600687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.600704 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.703923 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.704011 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.704049 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.704081 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.704100 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.736639 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.736708 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:38 crc kubenswrapper[4817]: E1128 14:30:38.736859 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:38 crc kubenswrapper[4817]: E1128 14:30:38.736959 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.807886 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.807963 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.807985 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.808019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.808043 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.914658 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.914748 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.914762 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.914815 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:38 crc kubenswrapper[4817]: I1128 14:30:38.914831 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:38Z","lastTransitionTime":"2025-11-28T14:30:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.018146 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.018214 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.018237 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.018268 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.018289 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.121147 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.121206 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.121218 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.121238 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.121252 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.223770 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.224216 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.224422 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.224607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.224832 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.328642 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.328839 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.328860 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.328882 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.328900 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.431877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.431934 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.431952 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.431977 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.431995 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.534956 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.535021 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.535042 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.535070 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.535089 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.637947 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.637988 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.637999 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.638016 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.638028 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.740839 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:39 crc kubenswrapper[4817]: E1128 14:30:39.740993 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.741253 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:39 crc kubenswrapper[4817]: E1128 14:30:39.741357 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.744129 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.744946 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.744988 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.745013 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.745034 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.848405 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.848465 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.848483 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.848506 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.848523 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.950917 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.950962 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.950979 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.950999 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:39 crc kubenswrapper[4817]: I1128 14:30:39.951013 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:39Z","lastTransitionTime":"2025-11-28T14:30:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.054160 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.054217 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.054261 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.054289 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.054306 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.157660 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.157763 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.157782 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.157807 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.157824 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.243669 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.243767 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.243786 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.243813 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.243831 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.266766 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:40Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.272867 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.272900 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.272913 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.272930 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.272941 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.289109 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:40Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.293993 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.294046 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.294064 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.294086 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.294103 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.344156 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:40Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.352758 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.352807 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.352822 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.352842 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.352858 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.371387 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:40Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.374649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.374683 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.374694 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.374710 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.374744 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.390967 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:40Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.391077 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.392550 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.392575 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.392592 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.392608 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.392619 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.494932 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.494990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.495005 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.495026 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.495042 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.599149 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.599217 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.599235 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.599263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.599282 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.702799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.702877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.702898 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.702923 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.702945 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.736545 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.736665 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.736768 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:40 crc kubenswrapper[4817]: E1128 14:30:40.736855 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.806857 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.806933 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.806952 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.806975 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.806991 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.910683 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.910792 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.910818 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.910851 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:40 crc kubenswrapper[4817]: I1128 14:30:40.910872 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:40Z","lastTransitionTime":"2025-11-28T14:30:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.014294 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.014366 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.014384 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.014409 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.014426 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.117578 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.117653 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.117693 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.117751 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.117776 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.220892 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.220941 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.220954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.220972 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.220986 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.323679 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.323732 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.323743 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.323759 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.323772 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.426533 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.426588 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.426605 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.426627 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.426644 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.529020 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.529108 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.529144 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.529174 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.529195 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.631587 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.631640 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.631657 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.631682 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.631701 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.734587 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.734714 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.734759 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.734782 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.734798 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.735918 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:41 crc kubenswrapper[4817]: E1128 14:30:41.736009 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.736069 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:41 crc kubenswrapper[4817]: E1128 14:30:41.736239 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.841689 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.841784 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.841809 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.841851 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.841878 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.944581 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.944644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.944656 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.944672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:41 crc kubenswrapper[4817]: I1128 14:30:41.944685 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:41Z","lastTransitionTime":"2025-11-28T14:30:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.047545 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.047577 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.047587 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.047602 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.047612 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.150178 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.150250 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.150263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.150295 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.150308 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.252988 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.253032 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.253043 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.253059 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.253074 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.355608 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.355647 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.355658 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.355675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.355686 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.458840 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.458898 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.458911 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.458934 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.458948 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.561588 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.561641 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.561652 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.561673 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.561685 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.664403 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.664452 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.664468 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.664493 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.664510 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.736525 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.736592 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:42 crc kubenswrapper[4817]: E1128 14:30:42.736662 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:42 crc kubenswrapper[4817]: E1128 14:30:42.736825 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.737798 4817 scope.go:117] "RemoveContainer" containerID="53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e" Nov 28 14:30:42 crc kubenswrapper[4817]: E1128 14:30:42.738016 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.766523 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.766551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.766562 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.766576 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.766589 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.870263 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.870333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.870354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.870390 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.870417 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.973802 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.973940 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.973962 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.973988 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:42 crc kubenswrapper[4817]: I1128 14:30:42.974044 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:42Z","lastTransitionTime":"2025-11-28T14:30:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.076606 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.076663 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.076680 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.076703 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.076752 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.180007 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.180049 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.180058 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.180074 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.180085 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.283182 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.283225 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.283234 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.283253 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.283262 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.387449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.387514 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.387527 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.387545 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.387565 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.490623 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.490661 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.490672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.490686 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.490696 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.494949 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:43 crc kubenswrapper[4817]: E1128 14:30:43.495222 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:43 crc kubenswrapper[4817]: E1128 14:30:43.495374 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:31:15.495337253 +0000 UTC m=+98.083315549 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.593044 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.593105 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.593127 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.593158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.593181 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.696968 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.697023 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.697040 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.697063 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.697079 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.736638 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.736698 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:43 crc kubenswrapper[4817]: E1128 14:30:43.736846 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:43 crc kubenswrapper[4817]: E1128 14:30:43.736980 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.799118 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.799185 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.799197 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.799213 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.799224 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.902413 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.902460 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.902476 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.902495 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:43 crc kubenswrapper[4817]: I1128 14:30:43.902506 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:43Z","lastTransitionTime":"2025-11-28T14:30:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.005574 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.005633 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.005649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.005672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.005689 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.109597 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.109639 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.109650 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.109666 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.109679 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.212560 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.212622 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.212638 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.212660 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.212681 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.315752 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.315789 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.315801 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.315817 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.315829 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.418301 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.418362 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.418387 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.418409 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.418426 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.521078 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.521118 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.521128 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.521142 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.521153 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.623600 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.623643 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.623655 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.623672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.623688 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.726138 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.726187 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.726204 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.726225 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.726240 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.735998 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:44 crc kubenswrapper[4817]: E1128 14:30:44.736078 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.736188 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:44 crc kubenswrapper[4817]: E1128 14:30:44.736233 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.828926 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.828967 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.828979 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.828995 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.829005 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.932185 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.932240 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.932251 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.932289 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:44 crc kubenswrapper[4817]: I1128 14:30:44.932304 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:44Z","lastTransitionTime":"2025-11-28T14:30:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.035235 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.035271 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.035281 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.035297 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.035308 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.138541 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.138635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.138654 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.138678 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.138698 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.193032 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/0.log" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.193085 4817 generic.go:334] "Generic (PLEG): container finished" podID="89086178-4127-42a3-8454-68e6d00c912a" containerID="e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5" exitCode=1 Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.193116 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerDied","Data":"e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.193514 4817 scope.go:117] "RemoveContainer" containerID="e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.213227 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.232266 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.241179 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.241218 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.241229 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.241249 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.241261 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.255689 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.271289 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.283518 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.292705 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.304651 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.321009 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.333533 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.343961 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.344214 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.344305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.344403 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.344478 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.347667 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.362589 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.381517 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.397968 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.414306 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.431323 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.447058 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.447085 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.447093 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.447105 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.447113 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.447795 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.460843 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.476498 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:45Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.549404 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.549456 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.549465 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.549480 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.549490 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.652187 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.652285 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.652313 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.652343 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.652365 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.736390 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:45 crc kubenswrapper[4817]: E1128 14:30:45.736573 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.736710 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:45 crc kubenswrapper[4817]: E1128 14:30:45.737001 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.754756 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.754799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.754811 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.754828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.754841 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.857027 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.857076 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.857092 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.857114 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.857132 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.961043 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.961082 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.961093 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.961109 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:45 crc kubenswrapper[4817]: I1128 14:30:45.961120 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:45Z","lastTransitionTime":"2025-11-28T14:30:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.064059 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.064109 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.064125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.064147 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.064163 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.167260 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.167317 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.167334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.167359 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.167380 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.198645 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/0.log" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.198708 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerStarted","Data":"58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.219653 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.235916 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.248744 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.262343 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.271036 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.271095 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.271112 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.271136 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.271153 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.280183 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.300746 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.315523 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.333814 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.360101 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.373522 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.373556 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.373566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.373580 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.373589 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.382487 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.391706 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.399671 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.408159 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.418496 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.430492 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.443303 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.455902 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.467888 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:46Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.475489 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.475528 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.475541 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.475557 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.475568 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.577966 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.578029 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.578047 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.578076 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.578095 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.680615 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.680672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.680688 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.680710 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.680749 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.736008 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:46 crc kubenswrapper[4817]: E1128 14:30:46.736189 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.736023 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:46 crc kubenswrapper[4817]: E1128 14:30:46.736514 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.782881 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.782946 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.783004 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.783032 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.783057 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.886039 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.886079 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.886088 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.886105 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.886117 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.988919 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.988958 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.988967 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.988980 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:46 crc kubenswrapper[4817]: I1128 14:30:46.988989 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:46Z","lastTransitionTime":"2025-11-28T14:30:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.091829 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.091893 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.091915 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.091943 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.091961 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.194876 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.194926 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.194943 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.194966 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.194982 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.297311 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.297376 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.297394 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.297416 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.297434 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.401159 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.401247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.401268 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.401291 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.401309 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.503189 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.503229 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.503237 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.503250 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.503261 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.605621 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.605907 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.605931 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.605945 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.605955 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.708591 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.708637 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.708653 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.708674 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.708690 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.736245 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.736312 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:47 crc kubenswrapper[4817]: E1128 14:30:47.736468 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:47 crc kubenswrapper[4817]: E1128 14:30:47.736571 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.767666 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.788135 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.803529 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.811397 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.811453 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.811472 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.811497 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.811513 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.820437 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.849393 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.863735 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.877281 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.890096 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.903825 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.913567 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.913619 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.913637 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.913658 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.913674 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:47Z","lastTransitionTime":"2025-11-28T14:30:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.926488 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.946231 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.967001 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.981944 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:47 crc kubenswrapper[4817]: I1128 14:30:47.994495 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:47Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.016098 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.016161 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.016173 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.016191 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.016204 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.018749 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:48Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.033026 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:48Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.049428 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:48Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.063432 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:48Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.118882 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.118945 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.118953 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.118968 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.118979 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.221334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.221397 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.221413 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.221431 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.221445 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.324601 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.324669 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.324694 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.324772 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.324799 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.427439 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.427526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.427551 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.427579 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.427602 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.530372 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.530424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.530436 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.530456 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.530468 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.634462 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.634547 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.634564 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.635046 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.635107 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.736688 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:48 crc kubenswrapper[4817]: E1128 14:30:48.736920 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.737077 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:48 crc kubenswrapper[4817]: E1128 14:30:48.737304 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.738794 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.738846 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.738863 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.738888 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.738908 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.842171 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.842298 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.842328 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.842355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.842376 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.944578 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.944645 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.944662 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.944687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:48 crc kubenswrapper[4817]: I1128 14:30:48.944704 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:48Z","lastTransitionTime":"2025-11-28T14:30:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.047876 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.047934 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.047950 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.047971 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.047990 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.151269 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.151338 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.151356 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.151382 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.151401 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.254600 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.254667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.254692 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.254754 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.254781 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.357812 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.357862 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.357877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.357902 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.357922 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.461344 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.461383 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.461394 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.461429 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.461440 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.563619 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.563674 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.563696 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.563751 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.563776 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.666690 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.666795 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.666819 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.666871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.666895 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.736785 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:49 crc kubenswrapper[4817]: E1128 14:30:49.736951 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.737105 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:49 crc kubenswrapper[4817]: E1128 14:30:49.737344 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.770008 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.770066 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.770083 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.770113 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.770131 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.873038 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.873124 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.873140 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.873167 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.873183 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.975089 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.975158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.975179 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.975585 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:49 crc kubenswrapper[4817]: I1128 14:30:49.975853 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:49Z","lastTransitionTime":"2025-11-28T14:30:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.078696 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.078754 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.078765 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.078782 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.078794 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.181711 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.181799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.181815 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.181841 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.181859 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.286112 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.286203 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.286220 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.286243 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.286260 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.389090 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.389131 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.389141 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.389156 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.389165 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.421822 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.421863 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.421876 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.421892 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.421903 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.441813 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:50Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.446378 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.446924 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.447111 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.447275 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.447411 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.468502 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:50Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.474277 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.474580 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.474765 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.474904 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.475038 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.494978 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:50Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.500905 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.500971 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.500982 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.501000 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.501012 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.521083 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:50Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.527311 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.527355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.527371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.527394 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.527410 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.550412 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:50Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.550769 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.553467 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.553516 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.553527 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.553545 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.553559 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.656659 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.656751 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.656770 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.656794 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.656811 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.736378 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.736467 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.736571 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:50 crc kubenswrapper[4817]: E1128 14:30:50.736665 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.759758 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.759833 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.759878 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.759913 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.759936 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.863262 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.863344 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.863368 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.863401 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.863423 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.965844 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.965916 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.965940 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.965972 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:50 crc kubenswrapper[4817]: I1128 14:30:50.965996 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:50Z","lastTransitionTime":"2025-11-28T14:30:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.068399 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.068424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.068431 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.068444 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.068493 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.171876 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.171932 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.171948 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.171974 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.171990 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.274669 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.274773 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.274798 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.274830 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.274852 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.377956 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.377994 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.378005 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.378021 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.378033 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.481061 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.481172 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.481192 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.481248 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.481267 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.583482 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.583542 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.583559 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.583582 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.583602 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.686476 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.686523 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.686534 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.686552 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.686565 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.735967 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.736065 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:51 crc kubenswrapper[4817]: E1128 14:30:51.736155 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:51 crc kubenswrapper[4817]: E1128 14:30:51.736301 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.789846 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.789884 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.789898 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.789916 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.789931 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.893689 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.893786 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.893809 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.893835 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.893856 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.996586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.996643 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.996659 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.996681 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:51 crc kubenswrapper[4817]: I1128 14:30:51.996697 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:51Z","lastTransitionTime":"2025-11-28T14:30:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.099474 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.099543 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.099566 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.099595 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.099620 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.203302 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.203427 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.203451 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.203488 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.203509 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.307158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.307201 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.307212 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.307230 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.307241 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.410655 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.410762 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.410781 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.410803 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.410816 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.513065 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.513109 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.513120 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.513136 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.513147 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.615524 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.615593 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.615611 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.615634 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.615651 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.719285 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.719354 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.719377 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.719404 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.719426 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.736205 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.736321 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:52 crc kubenswrapper[4817]: E1128 14:30:52.736526 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:52 crc kubenswrapper[4817]: E1128 14:30:52.736670 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.822592 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.822648 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.822665 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.822689 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.822706 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.925671 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.925770 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.925797 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.925824 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:52 crc kubenswrapper[4817]: I1128 14:30:52.925840 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:52Z","lastTransitionTime":"2025-11-28T14:30:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.028634 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.028698 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.028756 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.028797 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.028819 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.131499 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.131573 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.131595 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.131621 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.131639 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.235298 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.235374 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.235399 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.235428 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.235450 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.338119 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.338178 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.338202 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.338229 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.338249 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.440939 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.440996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.441012 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.441034 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.441051 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.543759 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.544675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.545004 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.545251 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.545430 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.649525 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.649589 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.649618 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.649649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.649672 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.736889 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.736902 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:53 crc kubenswrapper[4817]: E1128 14:30:53.737625 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:53 crc kubenswrapper[4817]: E1128 14:30:53.737818 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.751990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.752079 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.752100 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.752143 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.752176 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.855529 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.855591 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.855613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.855644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.855664 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.958487 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.958531 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.958548 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.958570 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:53 crc kubenswrapper[4817]: I1128 14:30:53.958585 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:53Z","lastTransitionTime":"2025-11-28T14:30:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.061298 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.061361 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.061385 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.061415 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.061437 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.164963 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.165045 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.165067 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.165095 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.165115 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.268218 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.268292 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.268315 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.268346 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.268368 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.371592 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.371693 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.371745 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.371778 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.371802 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.474692 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.474799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.474824 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.474871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.474896 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.578183 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.578250 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.578268 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.578291 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.578310 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.680556 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.680635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.680655 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.680681 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.680701 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.736108 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:54 crc kubenswrapper[4817]: E1128 14:30:54.736323 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.736468 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:54 crc kubenswrapper[4817]: E1128 14:30:54.736694 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.738192 4817 scope.go:117] "RemoveContainer" containerID="53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.782666 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.782714 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.782768 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.782788 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.782800 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.886758 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.886990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.887015 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.887045 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.887067 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.991487 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.991557 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.991579 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.991607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:54 crc kubenswrapper[4817]: I1128 14:30:54.991638 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:54Z","lastTransitionTime":"2025-11-28T14:30:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.095530 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.095590 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.095611 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.095638 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.095659 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.198671 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.198784 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.198817 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.198850 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.198870 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.237029 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/2.log" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.241351 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.241935 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.257514 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.271984 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.287547 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.301530 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.301575 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.301586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.301607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.301618 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.310311 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.326263 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.344367 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.365704 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.388019 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.399414 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.403599 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.403647 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.403658 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.403672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.403684 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.412922 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.424072 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.435334 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.454550 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.468390 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.481931 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.494580 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.506287 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.506334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.506348 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.506386 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.506400 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.508212 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.521945 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.609327 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.609362 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.609371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.609385 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.609397 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.712558 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.712594 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.712602 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.712620 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.712633 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.736802 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:55 crc kubenswrapper[4817]: E1128 14:30:55.736914 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.737020 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:55 crc kubenswrapper[4817]: E1128 14:30:55.737224 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.815605 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.815687 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.815712 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.815776 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.815797 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.919434 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.919490 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.919502 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.919520 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:55 crc kubenswrapper[4817]: I1128 14:30:55.919535 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:55Z","lastTransitionTime":"2025-11-28T14:30:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.022339 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.022411 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.022433 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.022462 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.022484 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.125556 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.125621 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.125638 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.125667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.125687 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.228978 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.229039 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.229080 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.229108 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.229126 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.253183 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/3.log" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.254482 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/2.log" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.260247 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" exitCode=1 Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.260317 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.260758 4817 scope.go:117] "RemoveContainer" containerID="53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.261591 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:30:56 crc kubenswrapper[4817]: E1128 14:30:56.261970 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.280415 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.296789 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.314419 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.331229 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.331271 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.331283 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.331299 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.331312 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.344033 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.365087 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.387923 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.409955 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.434159 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.434424 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.434581 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.434770 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.434942 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.442444 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c4c9535c821eadf610f5509dd6b149816a37d079ba27b0379c6dc88085b91e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:27Z\\\",\\\"message\\\":\\\" 6487 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1128 14:30:27.623437 6487 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.623334 6487 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 14:30:27.622712 6487 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8\\\\nI1128 14:30:27.6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:55Z\\\",\\\"message\\\":\\\"30:55.736833 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1128 14:30:55.736855 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1128 14:30:55.736861 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF1128 14:30:55.736874 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:55.736879 6834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.470288 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.487633 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.505090 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.523536 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.538464 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.538854 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.539072 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.539315 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.539504 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.543684 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.561367 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.581956 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.601138 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.621261 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.643280 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.643642 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.643823 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.643977 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.644106 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.644400 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:56Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.736137 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.736175 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:56 crc kubenswrapper[4817]: E1128 14:30:56.736312 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:56 crc kubenswrapper[4817]: E1128 14:30:56.736402 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.746789 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.746848 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.746871 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.746898 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.746920 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.850596 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.850659 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.850674 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.850696 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.850711 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.953526 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.953636 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.953653 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.953676 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:56 crc kubenswrapper[4817]: I1128 14:30:56.953691 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:56Z","lastTransitionTime":"2025-11-28T14:30:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.056644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.057139 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.057176 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.057214 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.057237 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.160397 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.160470 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.160488 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.160511 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.160529 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.263125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.263202 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.263225 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.263255 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.263281 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.267183 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/3.log" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.272947 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:30:57 crc kubenswrapper[4817]: E1128 14:30:57.273245 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.294267 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.315267 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.332752 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.350260 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.367401 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.367457 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.367474 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.367501 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.367520 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.368358 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.401408 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.422112 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.441910 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.462985 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.469882 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.469925 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.469941 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.469970 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.469996 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.491640 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:55Z\\\",\\\"message\\\":\\\"30:55.736833 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1128 14:30:55.736855 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1128 14:30:55.736861 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF1128 14:30:55.736874 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:55.736879 6834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.516830 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.535312 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.552915 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.572698 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.573389 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.573454 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.573472 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.573497 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.573516 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.595252 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.619782 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.640743 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.662499 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.680032 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.680090 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.680102 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.680122 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.680138 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.736328 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.736413 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:57 crc kubenswrapper[4817]: E1128 14:30:57.736487 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:57 crc kubenswrapper[4817]: E1128 14:30:57.736552 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.765484 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.780015 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.782843 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.782897 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.782918 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.782943 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.782961 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.799604 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.817040 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.837031 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.854291 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.873114 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.885598 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.885671 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.885684 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.885703 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.885748 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.893048 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.912834 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.944398 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:55Z\\\",\\\"message\\\":\\\"30:55.736833 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1128 14:30:55.736855 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1128 14:30:55.736861 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF1128 14:30:55.736874 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:55.736879 6834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.961974 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.976491 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.987707 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:57Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.988922 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.988978 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.988995 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.989026 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:57 crc kubenswrapper[4817]: I1128 14:30:57.989050 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:57Z","lastTransitionTime":"2025-11-28T14:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.003973 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.022990 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.038581 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.063982 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.078911 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:58Z is after 2025-08-24T17:21:41Z" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.092154 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.092192 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.092203 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.092220 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.092232 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.194552 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.194613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.194636 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.194663 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.194683 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.298286 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.298355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.298381 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.298411 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.298433 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.400667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.400709 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.400742 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.400757 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.400769 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.504032 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.504080 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.504091 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.504108 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.504120 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.606934 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.606990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.607007 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.607032 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.607050 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.710640 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.710800 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.710828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.710856 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.710876 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.735958 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.735999 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:30:58 crc kubenswrapper[4817]: E1128 14:30:58.736134 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:30:58 crc kubenswrapper[4817]: E1128 14:30:58.736389 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.813334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.813367 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.813378 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.813394 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.813406 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.917089 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.917145 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.917161 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.917183 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:58 crc kubenswrapper[4817]: I1128 14:30:58.917201 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:58Z","lastTransitionTime":"2025-11-28T14:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.020403 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.020448 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.020464 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.020486 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.020502 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.123837 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.123909 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.123931 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.123956 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.123973 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.226748 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.226806 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.226828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.226855 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.226879 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.329315 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.329386 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.329410 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.329438 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.329458 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.432701 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.432852 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.432879 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.432908 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.432928 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.536211 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.536295 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.536317 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.536343 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.536363 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.639219 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.639300 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.639323 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.639352 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.639373 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.736359 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.736504 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:30:59 crc kubenswrapper[4817]: E1128 14:30:59.736567 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:30:59 crc kubenswrapper[4817]: E1128 14:30:59.736712 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.742391 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.742472 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.742489 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.742513 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.742531 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.845509 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.845601 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.845622 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.845647 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.845667 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.948494 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.948546 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.948563 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.948584 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:30:59 crc kubenswrapper[4817]: I1128 14:30:59.948600 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:30:59Z","lastTransitionTime":"2025-11-28T14:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.051226 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.051555 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.051798 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.052014 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.052225 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.159990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.160084 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.160104 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.160128 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.160145 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.263600 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.263649 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.263667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.263689 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.263705 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.367530 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.367569 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.367604 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.367626 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.367640 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.470812 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.470859 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.470878 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.470900 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.470917 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574176 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574262 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574282 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574308 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574326 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574427 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574573 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574602 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.574570967 +0000 UTC m=+147.162549233 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574652 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574746 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574764 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574795 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.574801 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574815 4817 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574857 4817 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574884 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.574857795 +0000 UTC m=+147.162836111 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574939 4817 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574966 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.574924296 +0000 UTC m=+147.162902612 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574975 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574996 4817 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.574997 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.574981518 +0000 UTC m=+147.162959904 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.575013 4817 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.575052 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.575043099 +0000 UTC m=+147.163021475 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.667990 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.668053 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.668069 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.668091 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.668111 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.689357 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.694377 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.694440 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.694458 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.694485 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.694504 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.713794 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.719471 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.719531 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.719548 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.719574 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.719594 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.735947 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.736034 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.736159 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.736378 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.740422 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.745049 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.745112 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.745131 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.745155 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.745171 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.765795 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.771096 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.771140 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.771158 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.771182 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.771199 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.791218 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T14:31:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"603c0640-9dfd-46d2-9c89-a0eb080f9b75\\\",\\\"systemUUID\\\":\\\"14944962-313a-40b6-b900-0cbeaaf380e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:00Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:00 crc kubenswrapper[4817]: E1128 14:31:00.791433 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.793675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.793773 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.793794 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.793819 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.793837 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.896667 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.896769 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.896794 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.896822 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.896844 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.999498 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.999542 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.999564 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.999590 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:00 crc kubenswrapper[4817]: I1128 14:31:00.999610 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:00Z","lastTransitionTime":"2025-11-28T14:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.102179 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.102266 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.102283 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.102308 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.102326 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.205545 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.205612 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.205635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.205666 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.205687 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.308579 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.308628 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.308643 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.308665 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.308683 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.411972 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.412019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.412030 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.412047 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.412060 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.515121 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.515231 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.515259 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.515287 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.515309 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.619125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.619186 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.619209 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.619247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.619271 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.722837 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.723290 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.723309 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.723333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.723354 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.736260 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.736355 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:01 crc kubenswrapper[4817]: E1128 14:31:01.736469 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:01 crc kubenswrapper[4817]: E1128 14:31:01.736594 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.827084 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.827148 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.827159 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.827175 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.827186 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.930069 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.930117 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.930128 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.930143 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:01 crc kubenswrapper[4817]: I1128 14:31:01.930152 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:01Z","lastTransitionTime":"2025-11-28T14:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.033013 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.033075 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.033097 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.033125 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.033146 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.136288 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.136358 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.136382 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.136413 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.136429 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.239819 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.239879 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.239900 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.239932 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.239957 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.343780 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.343852 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.343877 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.343907 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.343930 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.447285 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.447344 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.447402 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.447434 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.447458 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.550455 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.550510 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.550528 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.550550 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.550567 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.653880 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.653938 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.653960 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.653988 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.654010 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.736518 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.736562 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:02 crc kubenswrapper[4817]: E1128 14:31:02.736691 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:02 crc kubenswrapper[4817]: E1128 14:31:02.736870 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.757015 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.757091 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.757109 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.757134 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.757151 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.859942 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.860022 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.860249 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.860274 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.860298 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.964010 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.964072 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.964089 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.964114 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:02 crc kubenswrapper[4817]: I1128 14:31:02.964131 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:02Z","lastTransitionTime":"2025-11-28T14:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.066793 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.066864 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.066886 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.066916 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.066934 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.169421 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.169477 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.169494 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.169518 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.169540 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.273228 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.273290 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.273308 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.273333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.273350 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.376553 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.376613 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.376635 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.376681 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.376708 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.480257 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.480323 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.480339 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.480364 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.480385 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.583716 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.583821 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.583842 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.583870 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.583887 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.686163 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.686213 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.686233 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.686258 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.686276 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.736202 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:03 crc kubenswrapper[4817]: E1128 14:31:03.736632 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.736372 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:03 crc kubenswrapper[4817]: E1128 14:31:03.736943 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.752399 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.788916 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.789302 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.789490 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.789672 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.789865 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.893343 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.894232 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.894377 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.894564 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.894759 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.997617 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.997692 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.997715 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.997766 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:03 crc kubenswrapper[4817]: I1128 14:31:03.997785 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:03Z","lastTransitionTime":"2025-11-28T14:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.101010 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.101096 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.101117 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.101144 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.101162 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.204283 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.204346 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.204363 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.204389 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.204408 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.307478 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.307544 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.307561 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.307582 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.307599 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.411084 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.411149 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.411166 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.411191 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.411209 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.514389 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.514449 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.514467 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.514490 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.514508 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.617120 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.617171 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.617187 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.617210 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.617227 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.720930 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.721001 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.721019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.721045 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.721062 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.736791 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.736885 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:04 crc kubenswrapper[4817]: E1128 14:31:04.737018 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:04 crc kubenswrapper[4817]: E1128 14:31:04.737112 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.824876 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.824941 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.824996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.825018 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.825037 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.928294 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.928348 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.928364 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.928385 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:04 crc kubenswrapper[4817]: I1128 14:31:04.928402 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:04Z","lastTransitionTime":"2025-11-28T14:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.032094 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.032151 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.032169 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.032193 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.032212 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.134525 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.134600 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.134618 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.134643 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.134661 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.237947 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.238019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.238036 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.238060 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.238082 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.340632 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.340713 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.340780 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.340810 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.340833 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.444089 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.444181 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.444221 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.444259 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.444287 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.547247 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.547312 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.547332 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.547359 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.547376 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.650030 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.650071 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.650083 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.650097 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.650109 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.737178 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.737227 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:05 crc kubenswrapper[4817]: E1128 14:31:05.737361 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:05 crc kubenswrapper[4817]: E1128 14:31:05.737517 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.752270 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.752345 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.752364 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.752388 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.752406 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.855224 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.855290 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.855312 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.855340 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.855364 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.958888 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.958945 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.958961 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.958984 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:05 crc kubenswrapper[4817]: I1128 14:31:05.959001 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:05Z","lastTransitionTime":"2025-11-28T14:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.061715 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.062181 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.062399 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.062617 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.062852 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.166051 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.166391 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.166565 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.166696 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.166950 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.270519 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.270612 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.270632 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.270661 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.270688 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.373792 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.373828 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.373841 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.373856 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.373865 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.476567 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.476651 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.476675 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.476706 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.476776 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.580327 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.580405 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.580433 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.580462 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.580483 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.683488 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.683555 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.683578 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.683606 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.683632 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.736819 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.736833 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:06 crc kubenswrapper[4817]: E1128 14:31:06.737064 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:06 crc kubenswrapper[4817]: E1128 14:31:06.737259 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.786598 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.786641 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.786658 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.786680 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.786698 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.889226 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.889549 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.889706 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.889922 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.890130 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.993873 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.993939 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.993960 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.993989 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:06 crc kubenswrapper[4817]: I1128 14:31:06.994013 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:06Z","lastTransitionTime":"2025-11-28T14:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.102631 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.102713 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.102773 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.102803 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.102842 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.206010 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.206099 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.206124 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.206157 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.206181 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.308833 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.308908 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.308928 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.308954 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.308974 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.412305 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.412363 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.412380 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.412404 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.412421 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.515275 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.515334 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.515355 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.515382 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.515405 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.618918 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.618996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.619019 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.619044 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.619061 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.722287 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.722360 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.722379 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.722404 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.722443 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.736884 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.736889 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:07 crc kubenswrapper[4817]: E1128 14:31:07.737119 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:07 crc kubenswrapper[4817]: E1128 14:31:07.737241 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.760743 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.779999 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.813874 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fa7491e-b213-4068-b2a8-77bf4979fc38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:55Z\\\",\\\"message\\\":\\\"30:55.736833 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1128 14:30:55.736855 6834 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1128 14:30:55.736861 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF1128 14:30:55.736874 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:30:55Z is after 2025-08-24T17:21:41Z]\\\\nI1128 14:30:55.736879 6834 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvq6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bktf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.825544 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.825606 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.825631 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.825660 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.825682 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.851843 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19f2800-93c0-4382-8a63-483f11e16632\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3b384bcdc7447bb76cd4cfe303fb77841f0b02435ea79b6beb901000321abc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f314a293dd7919797e36be72040556654fc572aa909fd7f3cd1456f98b36337f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://553bf4151976d2fe066270f6acfe4700c20009eea90de928ae6b6a45dc1bab32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e81d3fd6413c1612b8ae596374862b94c4054758ab45ed6283efcf501c1058af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d45733ae2ccdd9e9498eeca5cdc78903645058a6a0cabcf3d88edeb8bdfa06e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://498bcc569e228c7154ee2a7254936bdf72e4f0aa99feb0ae238d1cf0a6e4b248\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75afae876a068d24cd5c11f0a4795facc1182ebb5263f85ddf3dc2bfec093bea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:30:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:30:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtdv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx9ch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.870025 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5nszl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efa9319a-a051-418e-9d47-7720827d45fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc05b664d8d2afbd5a0b30b6670d8546775fd27e2569f606b1797b81b8d7b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-88mn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5nszl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.886190 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfd936ce-70b3-4e00-944f-f505238d1ab2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2ggdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zvnxb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.904238 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42c29379-2daf-4d13-b66a-3dedd8091d45\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://731b731437c883c75b164a1cbf1e7e8753729848c76ddeb2824702e5f2347065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba2c78d9c514b342908844ecafae363e252372c1e0524a574e4043eb2d63dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01b179fb18fefbe3512ba3a190e0aab51ca7b950c406eb57e92447796031a2cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17eb9813b0f54f277a92f8bd141e70f08b2e73c200b5144211ada641ec0f1697\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.919056 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1a02c3e-f459-4aef-bcf0-2522a9f5f34a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://933c4885ef9df2b3a4ca84ff7339657b932898469179395b4ba5140a248f9fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e50e590d43ee1c647bf22ff2b37fe9f4b14c7b4115161b6245208bde48073c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e50e590d43ee1c647bf22ff2b37fe9f4b14c7b4115161b6245208bde48073c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.929422 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.929473 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.929490 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.929512 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.929529 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:07Z","lastTransitionTime":"2025-11-28T14:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.939968 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sbhgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89086178-4127-42a3-8454-68e6d00c912a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T14:30:44Z\\\",\\\"message\\\":\\\"2025-11-28T14:29:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc\\\\n2025-11-28T14:29:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e8990c78-445d-454b-b30f-4f5e46ce97dc to /host/opt/cni/bin/\\\\n2025-11-28T14:29:59Z [verbose] multus-daemon started\\\\n2025-11-28T14:29:59Z [verbose] Readiness Indicator file check\\\\n2025-11-28T14:30:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sf9g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sbhgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.962299 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd1a8a66-95ea-4dc3-b906-8fb49bb61503\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T14:29:55Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 14:29:50.244952 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 14:29:50.246597 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-525630835/tls.crt::/tmp/serving-cert-525630835/tls.key\\\\\\\"\\\\nI1128 14:29:55.826746 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 14:29:55.832854 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 14:29:55.832895 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 14:29:55.832934 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 14:29:55.832964 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 14:29:55.844049 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 14:29:55.844131 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 14:29:55.844143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 14:29:55.844136 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 14:29:55.844152 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 14:29:55.844179 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 14:29:55.844186 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 14:29:55.844193 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 14:29:55.847201 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.979840 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15feee6093123961c3bff6dbef0e9895e46071fac4ad1c0a3d8141e215a53a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:07 crc kubenswrapper[4817]: I1128 14:31:07.996152 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"303f788a-102b-41e8-b4a8-66e69482a0fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11b2c31512f9f8f5a2f1e7ce4ec5e6d8dc88430f5e260eeac76d2d5378b83a2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ad5a47d1342a811a64a42d4d8f507e4e0222ccfbde11e98f372b4b88bbb9623\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a053cf3e8a566c5ab79e2e3cf1a8eaa06e32498e7001a475f5b2ad7eff643583\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:07Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.014133 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e6fe0bf6335c1041ec96a87415f9f688bb9e835620dadee462abd27cba4cf9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.032799 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.032869 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.032888 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.032914 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.032932 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.035815 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://371f5ddb42c924c03a2b33f8ffb26176d78ce27536379ff2cefe5a49100b1ab3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba419fe282917609f6a30d621157c7cecad0e9cede8bee4c08adaaac8bfe71c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.051100 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wn7r7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff6f26f-fe95-4e3f-a7d8-123897ca4bf6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe3241b145d7e3da7bc4c5082263abfbea175d32a1e41ee304b071911bc40b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bhts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wn7r7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.068973 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f904b2eaf4603bad548a8b1df8611175f8e08bf05de8a30c5d182ba569260c19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jsw76\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6kh49\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.086487 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d3ad39a-79e7-4547-a004-00e2826f1bd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:30:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://597d112fe47ebc1298dc71e9bfe327745ae9adcfa242be69f8e0a15b3e85f94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80679d53dad5ad162193ea7d3a61c330208ddfc09981a325e3c1285cc297b482\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:30:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgv7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:30:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pv4m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.118307 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b67d4-b830-4de5-85c1-89c1a88a82a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2867e1a9c6d8032eb54551629da52375aee1fe4ad80ce937ef7a26603075158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd7733a5a29dac1a6b20dc12af37df11a9a4f9b784df6ce23765bdbc8dd3a25c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d12fd53c2ced075f022802902e5852a65482ac0a6b2eb8d23f32a2637722e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac93204ebc9730f90cf9cd8ca510f272874144c44540b49e911f9322c85a0cdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8119b56ec63f2558a5602a5993908089632af1cb53827521e7834e859c85d247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T14:29:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0182528750fd0d3d919926f86de6bcd08716c9b205686158760ebe0797ded21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da993268fccaf7f3ce69d87a35145b4b4a17341c9115062849975bb640ba40c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b3c379e828fe02d85929d18ffb979d5665dd66e0897fd695a94e8b0184d4fba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T14:29:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T14:29:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T14:29:37Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.136916 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.136985 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.137007 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.137035 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.137059 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.139171 4817 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T14:29:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T14:31:08Z is after 2025-08-24T17:21:41Z" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.239421 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.239470 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.239479 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.239493 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.239503 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.343041 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.343133 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.343153 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.343185 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.343243 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.446272 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.446330 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.446346 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.446369 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.446387 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.548975 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.549052 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.549070 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.549097 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.549116 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.653553 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.653624 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.653644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.653677 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.653697 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.736597 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.736615 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:08 crc kubenswrapper[4817]: E1128 14:31:08.736818 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:08 crc kubenswrapper[4817]: E1128 14:31:08.736917 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.756949 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.757008 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.757025 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.757048 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.757066 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.861030 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.861111 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.861142 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.861177 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.861201 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.964505 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.964582 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.964610 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.964644 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:08 crc kubenswrapper[4817]: I1128 14:31:08.964667 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:08Z","lastTransitionTime":"2025-11-28T14:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.068166 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.068203 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.068211 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.068225 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.068234 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.171170 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.171246 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.171279 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.171311 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.171334 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.274929 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.274980 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.274992 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.275008 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.275024 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.378527 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.378586 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.378607 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.378632 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.378651 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.481713 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.481789 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.481801 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.481820 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.481831 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.583887 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.583937 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.583947 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.583965 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.583978 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.687242 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.687333 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.687353 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.687376 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.687394 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.736898 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:09 crc kubenswrapper[4817]: E1128 14:31:09.737144 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.737195 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:09 crc kubenswrapper[4817]: E1128 14:31:09.737659 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.790689 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.790780 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.790803 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.790826 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.790844 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.894156 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.894349 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.894371 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.894395 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.894414 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.998015 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.998090 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.998110 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.998137 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:09 crc kubenswrapper[4817]: I1128 14:31:09.998160 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:09Z","lastTransitionTime":"2025-11-28T14:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.100817 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.100873 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.100887 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.100909 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.100920 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.204816 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.204892 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.204912 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.204941 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.204959 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.307923 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.307978 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.307996 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.308020 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.308038 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.410864 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.410930 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.410948 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.410977 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.410994 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.514245 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.514307 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.514324 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.514347 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.514366 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.617819 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.617901 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.617944 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.617975 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.617994 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.720989 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.721050 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.721068 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.721092 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.721112 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.735899 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:10 crc kubenswrapper[4817]: E1128 14:31:10.736055 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.735899 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:10 crc kubenswrapper[4817]: E1128 14:31:10.736181 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.823620 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.823715 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.823795 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.823833 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.823870 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.926289 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.926352 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.926363 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.926381 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:10 crc kubenswrapper[4817]: I1128 14:31:10.926393 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:10Z","lastTransitionTime":"2025-11-28T14:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.029230 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.029277 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.029288 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.029306 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.029318 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:11Z","lastTransitionTime":"2025-11-28T14:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.044114 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.044181 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.044195 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.044211 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.044221 4817 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T14:31:11Z","lastTransitionTime":"2025-11-28T14:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.114796 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88"] Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.115765 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.118177 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.118554 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.120280 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.120347 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.158221 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=75.158184353 podStartE2EDuration="1m15.158184353s" podCreationTimestamp="2025-11-28 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.142688218 +0000 UTC m=+93.730666484" watchObservedRunningTime="2025-11-28 14:31:11.158184353 +0000 UTC m=+93.746162649" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.173667 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-sbhgx" podStartSLOduration=74.173639478 podStartE2EDuration="1m14.173639478s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.173354191 +0000 UTC m=+93.761332477" watchObservedRunningTime="2025-11-28 14:31:11.173639478 +0000 UTC m=+93.761617774" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.196254 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/07eeb5cf-9c42-4364-a31a-fbc883ae8930-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.196354 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07eeb5cf-9c42-4364-a31a-fbc883ae8930-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.196408 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/07eeb5cf-9c42-4364-a31a-fbc883ae8930-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.196444 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07eeb5cf-9c42-4364-a31a-fbc883ae8930-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.196573 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/07eeb5cf-9c42-4364-a31a-fbc883ae8930-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.228306 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.228257541 podStartE2EDuration="1m12.228257541s" podCreationTimestamp="2025-11-28 14:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.209615526 +0000 UTC m=+93.797593812" watchObservedRunningTime="2025-11-28 14:31:11.228257541 +0000 UTC m=+93.816235817" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.295183 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=73.295160878 podStartE2EDuration="1m13.295160878s" podCreationTimestamp="2025-11-28 14:29:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.290095979 +0000 UTC m=+93.878074255" watchObservedRunningTime="2025-11-28 14:31:11.295160878 +0000 UTC m=+93.883139164" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298166 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/07eeb5cf-9c42-4364-a31a-fbc883ae8930-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298224 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/07eeb5cf-9c42-4364-a31a-fbc883ae8930-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298253 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07eeb5cf-9c42-4364-a31a-fbc883ae8930-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298289 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/07eeb5cf-9c42-4364-a31a-fbc883ae8930-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298315 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07eeb5cf-9c42-4364-a31a-fbc883ae8930-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298346 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/07eeb5cf-9c42-4364-a31a-fbc883ae8930-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.298383 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/07eeb5cf-9c42-4364-a31a-fbc883ae8930-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.299473 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/07eeb5cf-9c42-4364-a31a-fbc883ae8930-service-ca\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.312860 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07eeb5cf-9c42-4364-a31a-fbc883ae8930-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.321538 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07eeb5cf-9c42-4364-a31a-fbc883ae8930-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-g5v88\" (UID: \"07eeb5cf-9c42-4364-a31a-fbc883ae8930\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.341458 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-wn7r7" podStartSLOduration=74.341439979 podStartE2EDuration="1m14.341439979s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.340673739 +0000 UTC m=+93.928651995" watchObservedRunningTime="2025-11-28 14:31:11.341439979 +0000 UTC m=+93.929418245" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.352576 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podStartSLOduration=74.352562203 podStartE2EDuration="1m14.352562203s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.352291436 +0000 UTC m=+93.940269702" watchObservedRunningTime="2025-11-28 14:31:11.352562203 +0000 UTC m=+93.940540469" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.365349 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pv4m8" podStartSLOduration=74.365325308 podStartE2EDuration="1m14.365325308s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.364845806 +0000 UTC m=+93.952824082" watchObservedRunningTime="2025-11-28 14:31:11.365325308 +0000 UTC m=+93.953303584" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.404207 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=8.40418991 podStartE2EDuration="8.40418991s" podCreationTimestamp="2025-11-28 14:31:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.402940318 +0000 UTC m=+93.990918584" watchObservedRunningTime="2025-11-28 14:31:11.40418991 +0000 UTC m=+93.992168176" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.404394 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.404390125 podStartE2EDuration="48.404390125s" podCreationTimestamp="2025-11-28 14:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.391374083 +0000 UTC m=+93.979352349" watchObservedRunningTime="2025-11-28 14:31:11.404390125 +0000 UTC m=+93.992368391" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.438059 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" Nov 28 14:31:11 crc kubenswrapper[4817]: W1128 14:31:11.458827 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07eeb5cf_9c42_4364_a31a_fbc883ae8930.slice/crio-aff66d9263d371585ab4d0220134aa20beb20181bb1d2736cbe48f219e3bdd04 WatchSource:0}: Error finding container aff66d9263d371585ab4d0220134aa20beb20181bb1d2736cbe48f219e3bdd04: Status 404 returned error can't find the container with id aff66d9263d371585ab4d0220134aa20beb20181bb1d2736cbe48f219e3bdd04 Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.478694 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rx9ch" podStartSLOduration=74.47867072 podStartE2EDuration="1m14.47867072s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.477066569 +0000 UTC m=+94.065044845" watchObservedRunningTime="2025-11-28 14:31:11.47867072 +0000 UTC m=+94.066649006" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.490820 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5nszl" podStartSLOduration=74.490798809 podStartE2EDuration="1m14.490798809s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:11.490082531 +0000 UTC m=+94.078060807" watchObservedRunningTime="2025-11-28 14:31:11.490798809 +0000 UTC m=+94.078777075" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.737067 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.737117 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:11 crc kubenswrapper[4817]: E1128 14:31:11.737581 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:11 crc kubenswrapper[4817]: E1128 14:31:11.737841 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:11 crc kubenswrapper[4817]: I1128 14:31:11.737998 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:31:11 crc kubenswrapper[4817]: E1128 14:31:11.738379 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:31:12 crc kubenswrapper[4817]: I1128 14:31:12.326792 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" event={"ID":"07eeb5cf-9c42-4364-a31a-fbc883ae8930","Type":"ContainerStarted","Data":"666f902e6d45e8e5de898ac37df28387311077b4d84a1acfd41109140be87d80"} Nov 28 14:31:12 crc kubenswrapper[4817]: I1128 14:31:12.326852 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" event={"ID":"07eeb5cf-9c42-4364-a31a-fbc883ae8930","Type":"ContainerStarted","Data":"aff66d9263d371585ab4d0220134aa20beb20181bb1d2736cbe48f219e3bdd04"} Nov 28 14:31:12 crc kubenswrapper[4817]: I1128 14:31:12.349189 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-g5v88" podStartSLOduration=75.349160039 podStartE2EDuration="1m15.349160039s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:12.348667326 +0000 UTC m=+94.936645612" watchObservedRunningTime="2025-11-28 14:31:12.349160039 +0000 UTC m=+94.937138365" Nov 28 14:31:12 crc kubenswrapper[4817]: I1128 14:31:12.736412 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:12 crc kubenswrapper[4817]: E1128 14:31:12.736974 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:12 crc kubenswrapper[4817]: I1128 14:31:12.736437 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:12 crc kubenswrapper[4817]: E1128 14:31:12.737255 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:13 crc kubenswrapper[4817]: I1128 14:31:13.736970 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:13 crc kubenswrapper[4817]: E1128 14:31:13.737141 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:13 crc kubenswrapper[4817]: I1128 14:31:13.737196 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:13 crc kubenswrapper[4817]: E1128 14:31:13.737377 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:14 crc kubenswrapper[4817]: I1128 14:31:14.736407 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:14 crc kubenswrapper[4817]: E1128 14:31:14.736618 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:14 crc kubenswrapper[4817]: I1128 14:31:14.736958 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:14 crc kubenswrapper[4817]: E1128 14:31:14.737150 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:15 crc kubenswrapper[4817]: I1128 14:31:15.545874 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:15 crc kubenswrapper[4817]: E1128 14:31:15.546119 4817 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:31:15 crc kubenswrapper[4817]: E1128 14:31:15.546497 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs podName:bfd936ce-70b3-4e00-944f-f505238d1ab2 nodeName:}" failed. No retries permitted until 2025-11-28 14:32:19.546464421 +0000 UTC m=+162.134442727 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs") pod "network-metrics-daemon-zvnxb" (UID: "bfd936ce-70b3-4e00-944f-f505238d1ab2") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 14:31:15 crc kubenswrapper[4817]: I1128 14:31:15.736020 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:15 crc kubenswrapper[4817]: E1128 14:31:15.736762 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:15 crc kubenswrapper[4817]: I1128 14:31:15.736896 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:15 crc kubenswrapper[4817]: E1128 14:31:15.737087 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:16 crc kubenswrapper[4817]: I1128 14:31:16.735960 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:16 crc kubenswrapper[4817]: I1128 14:31:16.736143 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:16 crc kubenswrapper[4817]: E1128 14:31:16.736486 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:16 crc kubenswrapper[4817]: E1128 14:31:16.736623 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:17 crc kubenswrapper[4817]: I1128 14:31:17.736572 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:17 crc kubenswrapper[4817]: I1128 14:31:17.736685 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:17 crc kubenswrapper[4817]: E1128 14:31:17.738418 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:17 crc kubenswrapper[4817]: E1128 14:31:17.738853 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:18 crc kubenswrapper[4817]: I1128 14:31:18.736080 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:18 crc kubenswrapper[4817]: I1128 14:31:18.736081 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:18 crc kubenswrapper[4817]: E1128 14:31:18.736406 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:18 crc kubenswrapper[4817]: E1128 14:31:18.736269 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:19 crc kubenswrapper[4817]: I1128 14:31:19.736658 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:19 crc kubenswrapper[4817]: E1128 14:31:19.737014 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:19 crc kubenswrapper[4817]: I1128 14:31:19.737059 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:19 crc kubenswrapper[4817]: E1128 14:31:19.737436 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:20 crc kubenswrapper[4817]: I1128 14:31:20.736172 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:20 crc kubenswrapper[4817]: I1128 14:31:20.736190 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:20 crc kubenswrapper[4817]: E1128 14:31:20.736523 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:20 crc kubenswrapper[4817]: E1128 14:31:20.736341 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:21 crc kubenswrapper[4817]: I1128 14:31:21.736318 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:21 crc kubenswrapper[4817]: I1128 14:31:21.736428 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:21 crc kubenswrapper[4817]: E1128 14:31:21.736551 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:21 crc kubenswrapper[4817]: E1128 14:31:21.736682 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:22 crc kubenswrapper[4817]: I1128 14:31:22.736226 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:22 crc kubenswrapper[4817]: I1128 14:31:22.736286 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:22 crc kubenswrapper[4817]: E1128 14:31:22.736370 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:22 crc kubenswrapper[4817]: E1128 14:31:22.736606 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:23 crc kubenswrapper[4817]: I1128 14:31:23.735885 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:23 crc kubenswrapper[4817]: I1128 14:31:23.735913 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:23 crc kubenswrapper[4817]: E1128 14:31:23.736027 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:23 crc kubenswrapper[4817]: E1128 14:31:23.736262 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:24 crc kubenswrapper[4817]: I1128 14:31:24.736963 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:24 crc kubenswrapper[4817]: I1128 14:31:24.737016 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:24 crc kubenswrapper[4817]: E1128 14:31:24.737495 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:24 crc kubenswrapper[4817]: E1128 14:31:24.737647 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:24 crc kubenswrapper[4817]: I1128 14:31:24.738943 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:31:24 crc kubenswrapper[4817]: E1128 14:31:24.739276 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:31:25 crc kubenswrapper[4817]: I1128 14:31:25.736669 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:25 crc kubenswrapper[4817]: I1128 14:31:25.736891 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:25 crc kubenswrapper[4817]: E1128 14:31:25.737009 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:25 crc kubenswrapper[4817]: E1128 14:31:25.737146 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:26 crc kubenswrapper[4817]: I1128 14:31:26.736594 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:26 crc kubenswrapper[4817]: E1128 14:31:26.736758 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:26 crc kubenswrapper[4817]: I1128 14:31:26.736772 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:26 crc kubenswrapper[4817]: E1128 14:31:26.737097 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:27 crc kubenswrapper[4817]: I1128 14:31:27.736452 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:27 crc kubenswrapper[4817]: E1128 14:31:27.738396 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:27 crc kubenswrapper[4817]: I1128 14:31:27.738509 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:27 crc kubenswrapper[4817]: E1128 14:31:27.738760 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:28 crc kubenswrapper[4817]: I1128 14:31:28.736952 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:28 crc kubenswrapper[4817]: I1128 14:31:28.736977 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:28 crc kubenswrapper[4817]: E1128 14:31:28.737130 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:28 crc kubenswrapper[4817]: E1128 14:31:28.737261 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:29 crc kubenswrapper[4817]: I1128 14:31:29.736555 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:29 crc kubenswrapper[4817]: E1128 14:31:29.737422 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:29 crc kubenswrapper[4817]: I1128 14:31:29.736576 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:29 crc kubenswrapper[4817]: E1128 14:31:29.737857 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:30 crc kubenswrapper[4817]: I1128 14:31:30.736965 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:30 crc kubenswrapper[4817]: I1128 14:31:30.736965 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:30 crc kubenswrapper[4817]: E1128 14:31:30.737191 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:30 crc kubenswrapper[4817]: E1128 14:31:30.737346 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.392803 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/1.log" Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.393714 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/0.log" Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.393848 4817 generic.go:334] "Generic (PLEG): container finished" podID="89086178-4127-42a3-8454-68e6d00c912a" containerID="58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf" exitCode=1 Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.393897 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerDied","Data":"58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf"} Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.393957 4817 scope.go:117] "RemoveContainer" containerID="e9f91be763cfc28be4702eb2cf77df84446abb9bde531e3ab2ca4fe1cada73f5" Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.394560 4817 scope.go:117] "RemoveContainer" containerID="58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf" Nov 28 14:31:31 crc kubenswrapper[4817]: E1128 14:31:31.394930 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-sbhgx_openshift-multus(89086178-4127-42a3-8454-68e6d00c912a)\"" pod="openshift-multus/multus-sbhgx" podUID="89086178-4127-42a3-8454-68e6d00c912a" Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.736671 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:31 crc kubenswrapper[4817]: I1128 14:31:31.736845 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:31 crc kubenswrapper[4817]: E1128 14:31:31.736965 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:31 crc kubenswrapper[4817]: E1128 14:31:31.737094 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:32 crc kubenswrapper[4817]: I1128 14:31:32.400711 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/1.log" Nov 28 14:31:32 crc kubenswrapper[4817]: I1128 14:31:32.736753 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:32 crc kubenswrapper[4817]: I1128 14:31:32.736786 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:32 crc kubenswrapper[4817]: E1128 14:31:32.736908 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:32 crc kubenswrapper[4817]: E1128 14:31:32.737069 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:33 crc kubenswrapper[4817]: I1128 14:31:33.736872 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:33 crc kubenswrapper[4817]: I1128 14:31:33.737027 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:33 crc kubenswrapper[4817]: E1128 14:31:33.737624 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:33 crc kubenswrapper[4817]: E1128 14:31:33.737904 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:34 crc kubenswrapper[4817]: I1128 14:31:34.736013 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:34 crc kubenswrapper[4817]: I1128 14:31:34.736084 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:34 crc kubenswrapper[4817]: E1128 14:31:34.736225 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:34 crc kubenswrapper[4817]: E1128 14:31:34.736345 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:35 crc kubenswrapper[4817]: I1128 14:31:35.736530 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:35 crc kubenswrapper[4817]: E1128 14:31:35.736711 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:35 crc kubenswrapper[4817]: I1128 14:31:35.737669 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:35 crc kubenswrapper[4817]: I1128 14:31:35.738006 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:31:35 crc kubenswrapper[4817]: E1128 14:31:35.738231 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bktf8_openshift-ovn-kubernetes(8fa7491e-b213-4068-b2a8-77bf4979fc38)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" Nov 28 14:31:35 crc kubenswrapper[4817]: E1128 14:31:35.738221 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:36 crc kubenswrapper[4817]: I1128 14:31:36.736086 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:36 crc kubenswrapper[4817]: I1128 14:31:36.736211 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:36 crc kubenswrapper[4817]: E1128 14:31:36.736245 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:36 crc kubenswrapper[4817]: E1128 14:31:36.736405 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:37 crc kubenswrapper[4817]: I1128 14:31:37.736256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:37 crc kubenswrapper[4817]: I1128 14:31:37.736357 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:37 crc kubenswrapper[4817]: E1128 14:31:37.739398 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:37 crc kubenswrapper[4817]: E1128 14:31:37.739622 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:37 crc kubenswrapper[4817]: E1128 14:31:37.771261 4817 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 28 14:31:37 crc kubenswrapper[4817]: E1128 14:31:37.840324 4817 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 14:31:38 crc kubenswrapper[4817]: I1128 14:31:38.736578 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:38 crc kubenswrapper[4817]: E1128 14:31:38.736770 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:38 crc kubenswrapper[4817]: I1128 14:31:38.736648 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:38 crc kubenswrapper[4817]: E1128 14:31:38.737224 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:39 crc kubenswrapper[4817]: I1128 14:31:39.736860 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:39 crc kubenswrapper[4817]: I1128 14:31:39.737008 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:39 crc kubenswrapper[4817]: E1128 14:31:39.737808 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:39 crc kubenswrapper[4817]: E1128 14:31:39.737933 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:40 crc kubenswrapper[4817]: I1128 14:31:40.736243 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:40 crc kubenswrapper[4817]: I1128 14:31:40.736292 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:40 crc kubenswrapper[4817]: E1128 14:31:40.736440 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:40 crc kubenswrapper[4817]: E1128 14:31:40.736569 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:41 crc kubenswrapper[4817]: I1128 14:31:41.736366 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:41 crc kubenswrapper[4817]: I1128 14:31:41.736581 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:41 crc kubenswrapper[4817]: I1128 14:31:41.736658 4817 scope.go:117] "RemoveContainer" containerID="58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf" Nov 28 14:31:41 crc kubenswrapper[4817]: E1128 14:31:41.737304 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:41 crc kubenswrapper[4817]: E1128 14:31:41.737408 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:42 crc kubenswrapper[4817]: I1128 14:31:42.448662 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/1.log" Nov 28 14:31:42 crc kubenswrapper[4817]: I1128 14:31:42.448804 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerStarted","Data":"c798287a6a545dc85c7121d38a6303407847b11a9390f3470f4f8f465b90166a"} Nov 28 14:31:42 crc kubenswrapper[4817]: I1128 14:31:42.736052 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:42 crc kubenswrapper[4817]: I1128 14:31:42.736125 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:42 crc kubenswrapper[4817]: E1128 14:31:42.736318 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:42 crc kubenswrapper[4817]: E1128 14:31:42.736493 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:42 crc kubenswrapper[4817]: E1128 14:31:42.842410 4817 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 14:31:43 crc kubenswrapper[4817]: I1128 14:31:43.736369 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:43 crc kubenswrapper[4817]: E1128 14:31:43.737161 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:43 crc kubenswrapper[4817]: I1128 14:31:43.736432 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:43 crc kubenswrapper[4817]: E1128 14:31:43.737682 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:44 crc kubenswrapper[4817]: I1128 14:31:44.736555 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:44 crc kubenswrapper[4817]: I1128 14:31:44.736631 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:44 crc kubenswrapper[4817]: E1128 14:31:44.736784 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:44 crc kubenswrapper[4817]: E1128 14:31:44.737049 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:45 crc kubenswrapper[4817]: I1128 14:31:45.736951 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:45 crc kubenswrapper[4817]: E1128 14:31:45.737147 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:45 crc kubenswrapper[4817]: I1128 14:31:45.737232 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:45 crc kubenswrapper[4817]: E1128 14:31:45.737367 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:46 crc kubenswrapper[4817]: I1128 14:31:46.736674 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:46 crc kubenswrapper[4817]: I1128 14:31:46.736809 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:46 crc kubenswrapper[4817]: E1128 14:31:46.736967 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:46 crc kubenswrapper[4817]: E1128 14:31:46.737460 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:46 crc kubenswrapper[4817]: I1128 14:31:46.737736 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.467550 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/3.log" Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.470356 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerStarted","Data":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.470850 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.511681 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podStartSLOduration=110.511668199 podStartE2EDuration="1m50.511668199s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:31:47.510894759 +0000 UTC m=+130.098873025" watchObservedRunningTime="2025-11-28 14:31:47.511668199 +0000 UTC m=+130.099646465" Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.690102 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zvnxb"] Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.690330 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:47 crc kubenswrapper[4817]: E1128 14:31:47.690689 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:47 crc kubenswrapper[4817]: I1128 14:31:47.736855 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:47 crc kubenswrapper[4817]: E1128 14:31:47.737919 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:47 crc kubenswrapper[4817]: E1128 14:31:47.843000 4817 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 14:31:48 crc kubenswrapper[4817]: I1128 14:31:48.736388 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:48 crc kubenswrapper[4817]: I1128 14:31:48.736401 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:48 crc kubenswrapper[4817]: E1128 14:31:48.737214 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:48 crc kubenswrapper[4817]: E1128 14:31:48.737388 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:49 crc kubenswrapper[4817]: I1128 14:31:49.736007 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:49 crc kubenswrapper[4817]: I1128 14:31:49.736359 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:49 crc kubenswrapper[4817]: E1128 14:31:49.736977 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:49 crc kubenswrapper[4817]: E1128 14:31:49.737162 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:50 crc kubenswrapper[4817]: I1128 14:31:50.735907 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:50 crc kubenswrapper[4817]: I1128 14:31:50.735962 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:50 crc kubenswrapper[4817]: E1128 14:31:50.736504 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:50 crc kubenswrapper[4817]: E1128 14:31:50.736358 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:51 crc kubenswrapper[4817]: I1128 14:31:51.736339 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:51 crc kubenswrapper[4817]: E1128 14:31:51.736509 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 14:31:51 crc kubenswrapper[4817]: I1128 14:31:51.736858 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:51 crc kubenswrapper[4817]: E1128 14:31:51.736983 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zvnxb" podUID="bfd936ce-70b3-4e00-944f-f505238d1ab2" Nov 28 14:31:52 crc kubenswrapper[4817]: I1128 14:31:52.736225 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:52 crc kubenswrapper[4817]: I1128 14:31:52.736232 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:52 crc kubenswrapper[4817]: E1128 14:31:52.736662 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 14:31:52 crc kubenswrapper[4817]: E1128 14:31:52.736863 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 14:31:53 crc kubenswrapper[4817]: I1128 14:31:53.736082 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:31:53 crc kubenswrapper[4817]: I1128 14:31:53.736096 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:31:53 crc kubenswrapper[4817]: I1128 14:31:53.738903 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 14:31:53 crc kubenswrapper[4817]: I1128 14:31:53.739187 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 14:31:53 crc kubenswrapper[4817]: I1128 14:31:53.739268 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 14:31:53 crc kubenswrapper[4817]: I1128 14:31:53.739229 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 14:31:54 crc kubenswrapper[4817]: I1128 14:31:54.736029 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:31:54 crc kubenswrapper[4817]: I1128 14:31:54.736062 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:31:54 crc kubenswrapper[4817]: I1128 14:31:54.739271 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 14:31:54 crc kubenswrapper[4817]: I1128 14:31:54.739589 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.673094 4817 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.714802 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.715629 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.715703 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-skq7g"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.716534 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.716989 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.717741 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.718093 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kzstg"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.718648 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.719098 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-v48xp"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.719970 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.721108 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.721633 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pk7bd"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.721765 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.721911 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.722049 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.722231 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.747270 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.747702 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.750183 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.750393 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.750529 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.750594 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.752124 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.752250 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.752433 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.752659 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.756303 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.757697 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.757698 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.759112 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.760636 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5kzwm"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.761214 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.763595 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.763678 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.763863 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767309 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767367 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767378 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767465 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767504 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767510 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767536 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767564 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767650 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767675 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767705 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767715 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767747 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767666 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767662 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767850 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767896 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767927 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767986 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.767673 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.770508 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jczpw"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.771030 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-9xjrk"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.771355 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.771487 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mxqrr"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.772048 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.772451 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.783575 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.809608 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.810434 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.811068 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.811297 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.811544 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.811654 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.812209 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.812231 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.812317 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.812335 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.812521 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.814008 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.815047 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.815181 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.817232 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.818845 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.819028 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.822243 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.831045 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v76hp"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.831459 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.831554 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.831804 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.831918 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832196 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832298 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832313 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832400 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832500 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832597 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832613 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832306 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832758 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832762 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832866 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832914 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832954 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.832600 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833137 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833169 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833242 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833255 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833417 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833449 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833583 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833046 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833052 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.834140 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833068 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.834273 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.834334 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833096 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.833104 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.834104 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.834823 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.836461 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-czfcx"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.836890 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.837224 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.837750 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.839575 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.840060 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.840360 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.840477 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q42hh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.841052 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.841080 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.841448 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.841968 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.842747 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.843514 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.843519 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.844503 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.845125 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.845586 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.846503 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.846920 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.847815 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.847924 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.853177 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.853358 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.853505 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.854847 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.855400 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.855737 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.855938 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.856291 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-skq7g"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.862622 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864081 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4552569-578c-4694-8e3b-1a3ec68380e9-serving-cert\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864115 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-config\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864141 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75f8d46c-6fda-4058-95bb-da00aadb176b-trusted-ca\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864163 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9cpz\" (UniqueName: \"kubernetes.io/projected/75f8d46c-6fda-4058-95bb-da00aadb176b-kube-api-access-w9cpz\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864188 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-client-ca\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864209 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-config\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864229 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trq9b\" (UniqueName: \"kubernetes.io/projected/a4552569-578c-4694-8e3b-1a3ec68380e9-kube-api-access-trq9b\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864255 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6pwd\" (UniqueName: \"kubernetes.io/projected/f890cb5e-2cff-4904-b9f5-3ecd702e869e-kube-api-access-b6pwd\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864275 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75f8d46c-6fda-4058-95bb-da00aadb176b-serving-cert\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864299 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-image-import-ca\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864320 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-trusted-ca-bundle\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864339 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-config\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864358 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864379 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864411 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-etcd-client\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864431 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-machine-approver-tls\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864454 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e20cea1b-1075-4334-a1b6-3634ee774393-config\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864474 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-config\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864495 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-audit\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864520 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-audit-policies\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864539 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-serving-cert\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864557 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-encryption-config\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864577 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-oauth-serving-cert\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864594 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-oauth-config\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864613 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99d03802-d8df-4c6d-9788-e125e4b3dc84-audit-dir\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864632 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-auth-proxy-config\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864653 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.864675 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.874054 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877077 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-etcd-client\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877121 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-etcd-serving-ca\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877146 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877183 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877209 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-encryption-config\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877231 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d8wm\" (UniqueName: \"kubernetes.io/projected/99d03802-d8df-4c6d-9788-e125e4b3dc84-kube-api-access-8d8wm\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877259 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-config\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877282 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877304 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877374 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e626351a-e0d2-48df-b470-e79a3f6f074e-node-pullsecrets\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877399 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e626351a-e0d2-48df-b470-e79a3f6f074e-audit-dir\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877418 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877478 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-ca\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877548 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-config\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877624 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-serving-cert\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877759 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.877800 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879060 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-client-ca\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879111 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-serving-cert\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879147 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-client\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879176 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-service-ca\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879209 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eed89a-6d2c-4b98-bf5b-13431b920d0c-serving-cert\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879244 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879271 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e20cea1b-1075-4334-a1b6-3634ee774393-images\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879330 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqxff\" (UniqueName: \"kubernetes.io/projected/e20cea1b-1075-4334-a1b6-3634ee774393-kube-api-access-fqxff\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879365 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krwn6\" (UniqueName: \"kubernetes.io/projected/8d7e9659-3a05-461d-ab42-ceda65eb40e9-kube-api-access-krwn6\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.881364 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.879417 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lkjz\" (UniqueName: \"kubernetes.io/projected/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-kube-api-access-2lkjz\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.886657 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-dir\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.887222 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e20cea1b-1075-4334-a1b6-3634ee774393-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.898640 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.898894 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.898965 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-service-ca\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.898992 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75f8d46c-6fda-4058-95bb-da00aadb176b-config\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899010 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899097 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-policies\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899135 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899154 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-serving-cert\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899179 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899195 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77prh\" (UniqueName: \"kubernetes.io/projected/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-kube-api-access-77prh\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899211 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwcfm\" (UniqueName: \"kubernetes.io/projected/67eed89a-6d2c-4b98-bf5b-13431b920d0c-kube-api-access-dwcfm\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899227 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85jbn\" (UniqueName: \"kubernetes.io/projected/8e94a88b-c6a1-4edd-8759-136275441876-kube-api-access-85jbn\") pod \"downloads-7954f5f757-5kzwm\" (UID: \"8e94a88b-c6a1-4edd-8759-136275441876\") " pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899253 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhjn5\" (UniqueName: \"kubernetes.io/projected/e626351a-e0d2-48df-b470-e79a3f6f074e-kube-api-access-lhjn5\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899373 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899607 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899654 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899791 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l6fvp"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.899963 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.900116 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.900813 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.900822 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.900843 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.901036 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4lbsj"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.901164 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.901661 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.901947 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.904575 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.904979 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.906116 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-82mjn"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.906759 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.907237 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.907318 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.907429 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.907457 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.908127 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.909090 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.910358 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.910986 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s56j8"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.912173 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.912774 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.913290 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4jlk5"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.914334 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.914572 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.916013 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.918063 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.918673 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.919748 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.922416 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.926005 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-mbk86"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.928254 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.928360 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-v48xp"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.929515 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kzstg"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.930592 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.932844 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.933078 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.938036 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.942862 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.946980 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9xjrk"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.947383 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5kzwm"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.950393 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jczpw"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.951367 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4jlk5"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.952770 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mxqrr"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.953414 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l6fvp"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.954694 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.955779 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-czfcx"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.956969 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.957622 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.958038 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.958907 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pk7bd"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.960517 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v76hp"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.961673 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.962583 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.964016 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.964728 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.965550 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.967526 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q42hh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.968516 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.969436 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.970612 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.971368 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.973241 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.973403 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zkksb"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.974925 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.975103 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.975347 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.976500 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s56j8"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.977491 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4lbsj"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.978542 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.978672 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mj798"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.980548 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.980633 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mbk86"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.980783 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.981630 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mj798"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.982553 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-z9v9l"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.983195 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.983503 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-z9v9l"] Nov 28 14:32:01 crc kubenswrapper[4817]: I1128 14:32:01.997985 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000001 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-oauth-config\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000028 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99d03802-d8df-4c6d-9788-e125e4b3dc84-audit-dir\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000046 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-auth-proxy-config\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000064 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000089 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000114 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000129 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-etcd-client\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000144 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-etcd-serving-ca\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000151 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99d03802-d8df-4c6d-9788-e125e4b3dc84-audit-dir\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000160 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000233 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000274 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-encryption-config\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000302 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d8wm\" (UniqueName: \"kubernetes.io/projected/99d03802-d8df-4c6d-9788-e125e4b3dc84-kube-api-access-8d8wm\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000330 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-config\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000356 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000396 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000426 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e626351a-e0d2-48df-b470-e79a3f6f074e-audit-dir\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000453 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000477 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e626351a-e0d2-48df-b470-e79a3f6f074e-node-pullsecrets\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000504 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-ca\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000527 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-config\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000553 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000580 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-serving-cert\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000646 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-client-ca\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000669 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-serving-cert\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000693 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-service-ca\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000713 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-client\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000764 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eed89a-6d2c-4b98-bf5b-13431b920d0c-serving-cert\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000790 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000815 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e20cea1b-1075-4334-a1b6-3634ee774393-images\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000841 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqxff\" (UniqueName: \"kubernetes.io/projected/e20cea1b-1075-4334-a1b6-3634ee774393-kube-api-access-fqxff\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000866 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krwn6\" (UniqueName: \"kubernetes.io/projected/8d7e9659-3a05-461d-ab42-ceda65eb40e9-kube-api-access-krwn6\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000894 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lkjz\" (UniqueName: \"kubernetes.io/projected/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-kube-api-access-2lkjz\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000917 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-dir\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000966 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-service-ca\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.000994 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75f8d46c-6fda-4058-95bb-da00aadb176b-config\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001018 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e20cea1b-1075-4334-a1b6-3634ee774393-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001059 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-policies\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001084 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001105 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-serving-cert\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001126 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001155 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwcfm\" (UniqueName: \"kubernetes.io/projected/67eed89a-6d2c-4b98-bf5b-13431b920d0c-kube-api-access-dwcfm\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001177 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85jbn\" (UniqueName: \"kubernetes.io/projected/8e94a88b-c6a1-4edd-8759-136275441876-kube-api-access-85jbn\") pod \"downloads-7954f5f757-5kzwm\" (UID: \"8e94a88b-c6a1-4edd-8759-136275441876\") " pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001202 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhjn5\" (UniqueName: \"kubernetes.io/projected/e626351a-e0d2-48df-b470-e79a3f6f074e-kube-api-access-lhjn5\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001227 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77prh\" (UniqueName: \"kubernetes.io/projected/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-kube-api-access-77prh\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001250 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4552569-578c-4694-8e3b-1a3ec68380e9-serving-cert\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001272 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-config\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001297 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75f8d46c-6fda-4058-95bb-da00aadb176b-trusted-ca\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001320 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9cpz\" (UniqueName: \"kubernetes.io/projected/75f8d46c-6fda-4058-95bb-da00aadb176b-kube-api-access-w9cpz\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001348 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-config\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001371 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-client-ca\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001396 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trq9b\" (UniqueName: \"kubernetes.io/projected/a4552569-578c-4694-8e3b-1a3ec68380e9-kube-api-access-trq9b\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001421 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6pwd\" (UniqueName: \"kubernetes.io/projected/f890cb5e-2cff-4904-b9f5-3ecd702e869e-kube-api-access-b6pwd\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001448 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75f8d46c-6fda-4058-95bb-da00aadb176b-serving-cert\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001453 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-auth-proxy-config\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001481 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-image-import-ca\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001508 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-trusted-ca-bundle\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001534 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-config\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001563 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001587 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001629 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-etcd-client\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001654 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-machine-approver-tls\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001681 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e20cea1b-1075-4334-a1b6-3634ee774393-config\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001709 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-config\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001751 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-audit\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001790 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-audit-policies\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001812 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-serving-cert\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001834 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-encryption-config\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.001857 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-oauth-serving-cert\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.002280 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-etcd-serving-ca\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.002472 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-dir\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.002822 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.002834 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-oauth-serving-cert\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.003187 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-service-ca\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.003356 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.003445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-config\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.004077 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75f8d46c-6fda-4058-95bb-da00aadb176b-config\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.004289 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-client-ca\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.004971 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-config\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.005241 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.005300 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e626351a-e0d2-48df-b470-e79a3f6f074e-audit-dir\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.005879 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-policies\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.006050 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e626351a-e0d2-48df-b470-e79a3f6f074e-node-pullsecrets\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.006605 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.006749 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-ca\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.006845 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-etcd-client\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.006855 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.006893 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-audit\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.007786 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99d03802-d8df-4c6d-9788-e125e4b3dc84-audit-policies\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.007864 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-image-import-ca\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.007921 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-client-ca\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.008086 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e20cea1b-1075-4334-a1b6-3634ee774393-config\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.008603 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-config\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.008800 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.009106 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-trusted-ca-bundle\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.009299 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.009364 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-service-ca\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.009471 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-config\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.009808 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-serving-cert\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.010402 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-etcd-client\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.010662 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-config\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.010818 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-encryption-config\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.011181 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-config\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.011321 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e20cea1b-1075-4334-a1b6-3634ee774393-images\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.011701 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75f8d46c-6fda-4058-95bb-da00aadb176b-trusted-ca\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.011987 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.012170 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e626351a-e0d2-48df-b470-e79a3f6f074e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.012494 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-etcd-client\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.012507 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eed89a-6d2c-4b98-bf5b-13431b920d0c-serving-cert\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.012900 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-oauth-config\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.013950 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-encryption-config\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014026 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014085 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e626351a-e0d2-48df-b470-e79a3f6f074e-serving-cert\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014462 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99d03802-d8df-4c6d-9788-e125e4b3dc84-serving-cert\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014544 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-machine-approver-tls\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014652 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014620 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014514 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75f8d46c-6fda-4058-95bb-da00aadb176b-serving-cert\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.014751 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e20cea1b-1075-4334-a1b6-3634ee774393-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.016355 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4552569-578c-4694-8e3b-1a3ec68380e9-serving-cert\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.018332 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.018415 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-serving-cert\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.021645 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.023512 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.026264 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.038493 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.069916 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.077797 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.096962 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.117400 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.138586 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.158261 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.177878 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.198405 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.219594 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.239268 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.260262 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.278771 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.298337 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.317194 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.337700 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.358126 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.378438 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.399032 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.418575 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.438301 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.461563 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.478316 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.498282 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.518852 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.538313 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.559152 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.578429 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.599119 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.618698 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.638797 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.657868 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.678889 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.698945 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.718697 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.738515 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.759075 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.778376 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.799229 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.818985 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.838907 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.856467 4817 request.go:700] Waited for 1.000353578s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackageserver-service-cert&limit=500&resourceVersion=0 Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.858571 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.878976 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.898844 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.918937 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.938186 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.958578 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 14:32:02 crc kubenswrapper[4817]: I1128 14:32:02.978790 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.009283 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.018630 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.039580 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.058984 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.078977 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.098582 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.118218 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.138345 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.159113 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.178301 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.199784 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.219199 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.238482 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.299333 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.318592 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.338604 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.358802 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.382683 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.397790 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.418414 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.444929 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.458982 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.477705 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.498625 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.518452 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.538530 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.558374 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.578061 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.598262 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.617778 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.638443 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.658540 4817 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.678657 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.698752 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.718305 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.739610 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.757998 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.798791 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqxff\" (UniqueName: \"kubernetes.io/projected/e20cea1b-1075-4334-a1b6-3634ee774393-kube-api-access-fqxff\") pod \"machine-api-operator-5694c8668f-skq7g\" (UID: \"e20cea1b-1075-4334-a1b6-3634ee774393\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.821532 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9cpz\" (UniqueName: \"kubernetes.io/projected/75f8d46c-6fda-4058-95bb-da00aadb176b-kube-api-access-w9cpz\") pod \"console-operator-58897d9998-pk7bd\" (UID: \"75f8d46c-6fda-4058-95bb-da00aadb176b\") " pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.846698 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krwn6\" (UniqueName: \"kubernetes.io/projected/8d7e9659-3a05-461d-ab42-ceda65eb40e9-kube-api-access-krwn6\") pod \"console-f9d7485db-9xjrk\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.850883 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.856809 4817 request.go:700] Waited for 1.852381034s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/serviceaccounts/openshift-controller-manager-sa/token Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.860935 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lkjz\" (UniqueName: \"kubernetes.io/projected/9bad542b-4641-4db9-9a0e-bc1f80fd0a32-kube-api-access-2lkjz\") pod \"machine-approver-56656f9798-ffg76\" (UID: \"9bad542b-4641-4db9-9a0e-bc1f80fd0a32\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.864932 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.889441 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trq9b\" (UniqueName: \"kubernetes.io/projected/a4552569-578c-4694-8e3b-1a3ec68380e9-kube-api-access-trq9b\") pod \"controller-manager-879f6c89f-kzstg\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.901855 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6pwd\" (UniqueName: \"kubernetes.io/projected/f890cb5e-2cff-4904-b9f5-3ecd702e869e-kube-api-access-b6pwd\") pod \"oauth-openshift-558db77b4-jczpw\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.911854 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d8wm\" (UniqueName: \"kubernetes.io/projected/99d03802-d8df-4c6d-9788-e125e4b3dc84-kube-api-access-8d8wm\") pod \"apiserver-7bbb656c7d-27xnz\" (UID: \"99d03802-d8df-4c6d-9788-e125e4b3dc84\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.912938 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.933063 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhjn5\" (UniqueName: \"kubernetes.io/projected/e626351a-e0d2-48df-b470-e79a3f6f074e-kube-api-access-lhjn5\") pod \"apiserver-76f77b778f-v48xp\" (UID: \"e626351a-e0d2-48df-b470-e79a3f6f074e\") " pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.942794 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.963859 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwcfm\" (UniqueName: \"kubernetes.io/projected/67eed89a-6d2c-4b98-bf5b-13431b920d0c-kube-api-access-dwcfm\") pod \"route-controller-manager-6576b87f9c-tfg8d\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.965638 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.981750 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77prh\" (UniqueName: \"kubernetes.io/projected/7b5b1a61-6e93-4e2c-b317-ac90df8381aa-kube-api-access-77prh\") pod \"etcd-operator-b45778765-mxqrr\" (UID: \"7b5b1a61-6e93-4e2c-b317-ac90df8381aa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:03 crc kubenswrapper[4817]: I1128 14:32:03.996597 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85jbn\" (UniqueName: \"kubernetes.io/projected/8e94a88b-c6a1-4edd-8759-136275441876-kube-api-access-85jbn\") pod \"downloads-7954f5f757-5kzwm\" (UID: \"8e94a88b-c6a1-4edd-8759-136275441876\") " pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.004006 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.022358 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034329 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5e5e62f0-77fd-446f-8166-154f715cea66-images\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034362 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034385 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5603525a-e561-4243-90ae-eb910aa81d29-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034400 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/db5b0254-7015-4320-b782-ac764446265f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034434 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-tls\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034458 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/59925674-795d-4e20-b71b-351cf566550d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4lbsj\" (UID: \"59925674-795d-4e20-b71b-351cf566550d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034473 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-service-ca-bundle\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034490 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcn7d\" (UniqueName: \"kubernetes.io/projected/55dbc750-3c63-44f9-ada1-aa282adb3562-kube-api-access-bcn7d\") pod \"control-plane-machine-set-operator-78cbb6b69f-kvssf\" (UID: \"55dbc750-3c63-44f9-ada1-aa282adb3562\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034506 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034533 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sg7l\" (UniqueName: \"kubernetes.io/projected/8b50b274-12b2-4ed3-ad0a-81365aeff72c-kube-api-access-6sg7l\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034555 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8whv\" (UniqueName: \"kubernetes.io/projected/9f893fe9-73c9-4758-a561-617a0afee23c-kube-api-access-q8whv\") pod \"cluster-samples-operator-665b6dd947-fwxdh\" (UID: \"9f893fe9-73c9-4758-a561-617a0afee23c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034571 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96abee8c-2ffa-4329-9f61-03e783f933b2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034589 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz9qp\" (UniqueName: \"kubernetes.io/projected/59925674-795d-4e20-b71b-351cf566550d-kube-api-access-tz9qp\") pod \"multus-admission-controller-857f4d67dd-4lbsj\" (UID: \"59925674-795d-4e20-b71b-351cf566550d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034603 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74f0b1b5-cf28-4dd7-9701-b7361caa5442-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034620 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034635 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2fqh\" (UniqueName: \"kubernetes.io/projected/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-kube-api-access-n2fqh\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034651 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wm9p\" (UniqueName: \"kubernetes.io/projected/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-kube-api-access-6wm9p\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034675 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdsbt\" (UniqueName: \"kubernetes.io/projected/1621fdc1-9b5b-4022-940a-7b8119be5b9d-kube-api-access-xdsbt\") pod \"package-server-manager-789f6589d5-cx5k6\" (UID: \"1621fdc1-9b5b-4022-940a-7b8119be5b9d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034821 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cc497af5-6bad-4dc2-a1a0-e574992346a5-metrics-tls\") pod \"dns-operator-744455d44c-q42hh\" (UID: \"cc497af5-6bad-4dc2-a1a0-e574992346a5\") " pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034843 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-stats-auth\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034871 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-bound-sa-token\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034894 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-config\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034910 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-secret-volume\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034934 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdmmj\" (UniqueName: \"kubernetes.io/projected/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-kube-api-access-pdmmj\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034958 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3fe840a-b55a-459a-98eb-f3263afd9184-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.034973 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035004 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035020 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/55dbc750-3c63-44f9-ada1-aa282adb3562-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-kvssf\" (UID: \"55dbc750-3c63-44f9-ada1-aa282adb3562\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035088 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f893fe9-73c9-4758-a561-617a0afee23c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fwxdh\" (UID: \"9f893fe9-73c9-4758-a561-617a0afee23c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035104 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjzts\" (UniqueName: \"kubernetes.io/projected/d7975b7a-932a-41de-91aa-06a5b67833a2-kube-api-access-xjzts\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035118 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3fe840a-b55a-459a-98eb-f3263afd9184-metrics-tls\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035132 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b50b274-12b2-4ed3-ad0a-81365aeff72c-serving-cert\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035147 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96abee8c-2ffa-4329-9f61-03e783f933b2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035161 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-certificates\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035177 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-metrics-certs\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035209 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sqsh\" (UniqueName: \"kubernetes.io/projected/96abee8c-2ffa-4329-9f61-03e783f933b2-kube-api-access-8sqsh\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035223 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e5e62f0-77fd-446f-8166-154f715cea66-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035246 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92c2898e-8e5b-4d26-b90f-97b357f5de80-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035277 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-config-volume\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035292 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e5e62f0-77fd-446f-8166-154f715cea66-proxy-tls\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035305 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/db5b0254-7015-4320-b782-ac764446265f-srv-cert\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035320 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxmvv\" (UniqueName: \"kubernetes.io/projected/92c2898e-8e5b-4d26-b90f-97b357f5de80-kube-api-access-gxmvv\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035336 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5603525a-e561-4243-90ae-eb910aa81d29-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035351 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7248143-1eb1-4858-a4a3-63d74fa4a94b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035365 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7248143-1eb1-4858-a4a3-63d74fa4a94b-config\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035381 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d845\" (UniqueName: \"kubernetes.io/projected/cc497af5-6bad-4dc2-a1a0-e574992346a5-kube-api-access-5d845\") pod \"dns-operator-744455d44c-q42hh\" (UID: \"cc497af5-6bad-4dc2-a1a0-e574992346a5\") " pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035395 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m255s\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-kube-api-access-m255s\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035426 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5603525a-e561-4243-90ae-eb910aa81d29-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035441 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8b50b274-12b2-4ed3-ad0a-81365aeff72c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035457 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-default-certificate\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035480 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-trusted-ca\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035495 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0f1f4b07-67d3-4920-8984-410096b9fe3d-proxy-tls\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035520 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74f0b1b5-cf28-4dd7-9701-b7361caa5442-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035534 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035559 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0f1f4b07-67d3-4920-8984-410096b9fe3d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035574 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7248143-1eb1-4858-a4a3-63d74fa4a94b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035599 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchd9\" (UniqueName: \"kubernetes.io/projected/797dc8be-e15a-49f7-b669-fecb0809759c-kube-api-access-zchd9\") pod \"migrator-59844c95c7-59g9q\" (UID: \"797dc8be-e15a-49f7-b669-fecb0809759c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035615 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvf8s\" (UniqueName: \"kubernetes.io/projected/db5b0254-7015-4320-b782-ac764446265f-kube-api-access-tvf8s\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035628 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f0b1b5-cf28-4dd7-9701-b7361caa5442-config\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035662 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qb7v\" (UniqueName: \"kubernetes.io/projected/5e5e62f0-77fd-446f-8166-154f715cea66-kube-api-access-5qb7v\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035684 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c2898e-8e5b-4d26-b90f-97b357f5de80-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035699 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sqsk\" (UniqueName: \"kubernetes.io/projected/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-kube-api-access-8sqsk\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035731 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f0ef5080-e168-4e13-92d5-8a41f51610da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035746 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035761 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3fe840a-b55a-459a-98eb-f3263afd9184-trusted-ca\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035797 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f0ef5080-e168-4e13-92d5-8a41f51610da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035819 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035835 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwwfg\" (UniqueName: \"kubernetes.io/projected/6b42da42-0ae3-43d1-be76-e400b3a124ac-kube-api-access-xwwfg\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035851 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d7975b7a-932a-41de-91aa-06a5b67833a2-tmpfs\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035901 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035918 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7975b7a-932a-41de-91aa-06a5b67833a2-webhook-cert\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035934 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/96abee8c-2ffa-4329-9f61-03e783f933b2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035951 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7975b7a-932a-41de-91aa-06a5b67833a2-apiservice-cert\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035968 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgnmc\" (UniqueName: \"kubernetes.io/projected/c3fe840a-b55a-459a-98eb-f3263afd9184-kube-api-access-lgnmc\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035983 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkn9k\" (UniqueName: \"kubernetes.io/projected/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-kube-api-access-dkn9k\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.035999 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1621fdc1-9b5b-4022-940a-7b8119be5b9d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cx5k6\" (UID: \"1621fdc1-9b5b-4022-940a-7b8119be5b9d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.036014 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47b29\" (UniqueName: \"kubernetes.io/projected/0f1f4b07-67d3-4920-8984-410096b9fe3d-kube-api-access-47b29\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.036040 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b42da42-0ae3-43d1-be76-e400b3a124ac-serving-cert\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.039937 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.539925125 +0000 UTC m=+147.127903391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.039958 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.055407 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.070266 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-skq7g"] Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.117894 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode20cea1b_1075_4334_a1b6_3634ee774393.slice/crio-eba9479662a56c4be73ec7fc2ca0c9b1752df9657910e871cce4105d7e26dec1 WatchSource:0}: Error finding container eba9479662a56c4be73ec7fc2ca0c9b1752df9657910e871cce4105d7e26dec1: Status 404 returned error can't find the container with id eba9479662a56c4be73ec7fc2ca0c9b1752df9657910e871cce4105d7e26dec1 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137004 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137143 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-tls\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.137179 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.63714605 +0000 UTC m=+147.225124326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137231 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/59925674-795d-4e20-b71b-351cf566550d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4lbsj\" (UID: \"59925674-795d-4e20-b71b-351cf566550d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137282 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-service-ca-bundle\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137336 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq6cr\" (UniqueName: \"kubernetes.io/projected/7bc1102b-0f53-4b51-b5c4-eeb118249dea-kube-api-access-kq6cr\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137367 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcn7d\" (UniqueName: \"kubernetes.io/projected/55dbc750-3c63-44f9-ada1-aa282adb3562-kube-api-access-bcn7d\") pod \"control-plane-machine-set-operator-78cbb6b69f-kvssf\" (UID: \"55dbc750-3c63-44f9-ada1-aa282adb3562\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137393 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137440 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sg7l\" (UniqueName: \"kubernetes.io/projected/8b50b274-12b2-4ed3-ad0a-81365aeff72c-kube-api-access-6sg7l\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137467 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e66dca3-d714-48b7-aceb-b9a826efcfdc-cert\") pod \"ingress-canary-z9v9l\" (UID: \"2e66dca3-d714-48b7-aceb-b9a826efcfdc\") " pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137495 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz9qp\" (UniqueName: \"kubernetes.io/projected/59925674-795d-4e20-b71b-351cf566550d-kube-api-access-tz9qp\") pod \"multus-admission-controller-857f4d67dd-4lbsj\" (UID: \"59925674-795d-4e20-b71b-351cf566550d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137521 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74f0b1b5-cf28-4dd7-9701-b7361caa5442-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137548 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137584 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8whv\" (UniqueName: \"kubernetes.io/projected/9f893fe9-73c9-4758-a561-617a0afee23c-kube-api-access-q8whv\") pod \"cluster-samples-operator-665b6dd947-fwxdh\" (UID: \"9f893fe9-73c9-4758-a561-617a0afee23c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96abee8c-2ffa-4329-9f61-03e783f933b2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137630 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2fqh\" (UniqueName: \"kubernetes.io/projected/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-kube-api-access-n2fqh\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137654 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wm9p\" (UniqueName: \"kubernetes.io/projected/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-kube-api-access-6wm9p\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137675 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdsbt\" (UniqueName: \"kubernetes.io/projected/1621fdc1-9b5b-4022-940a-7b8119be5b9d-kube-api-access-xdsbt\") pod \"package-server-manager-789f6589d5-cx5k6\" (UID: \"1621fdc1-9b5b-4022-940a-7b8119be5b9d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137696 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/87433151-5993-4f35-a3f3-5b39e3186ff5-signing-key\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137805 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-plugins-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137843 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cc497af5-6bad-4dc2-a1a0-e574992346a5-metrics-tls\") pod \"dns-operator-744455d44c-q42hh\" (UID: \"cc497af5-6bad-4dc2-a1a0-e574992346a5\") " pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137866 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-srv-cert\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137885 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-stats-auth\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137908 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-bound-sa-token\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137956 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-config\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.137979 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-secret-volume\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138013 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdmmj\" (UniqueName: \"kubernetes.io/projected/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-kube-api-access-pdmmj\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138049 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3fe840a-b55a-459a-98eb-f3263afd9184-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138109 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl5jl\" (UniqueName: \"kubernetes.io/projected/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-kube-api-access-cl5jl\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138147 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138171 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3870a07-6c0e-4df8-8d96-159600ab5bd1-config\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138209 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138235 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/55dbc750-3c63-44f9-ada1-aa282adb3562-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-kvssf\" (UID: \"55dbc750-3c63-44f9-ada1-aa282adb3562\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138264 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbtgm\" (UniqueName: \"kubernetes.io/projected/f269f51f-578d-49f0-8696-c2c69e8763fd-kube-api-access-kbtgm\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138292 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f893fe9-73c9-4758-a561-617a0afee23c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fwxdh\" (UID: \"9f893fe9-73c9-4758-a561-617a0afee23c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138316 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjzts\" (UniqueName: \"kubernetes.io/projected/d7975b7a-932a-41de-91aa-06a5b67833a2-kube-api-access-xjzts\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138343 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b50b274-12b2-4ed3-ad0a-81365aeff72c-serving-cert\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138364 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3fe840a-b55a-459a-98eb-f3263afd9184-metrics-tls\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138388 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/87433151-5993-4f35-a3f3-5b39e3186ff5-signing-cabundle\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138411 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96abee8c-2ffa-4329-9f61-03e783f933b2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138446 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-certificates\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138467 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9znlm\" (UniqueName: \"kubernetes.io/projected/2e66dca3-d714-48b7-aceb-b9a826efcfdc-kube-api-access-9znlm\") pod \"ingress-canary-z9v9l\" (UID: \"2e66dca3-d714-48b7-aceb-b9a826efcfdc\") " pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138491 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-config-volume\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138527 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-metrics-certs\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138553 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sqsh\" (UniqueName: \"kubernetes.io/projected/96abee8c-2ffa-4329-9f61-03e783f933b2-kube-api-access-8sqsh\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138576 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e5e62f0-77fd-446f-8166-154f715cea66-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138599 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f269f51f-578d-49f0-8696-c2c69e8763fd-node-bootstrap-token\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138631 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpn2f\" (UniqueName: \"kubernetes.io/projected/87433151-5993-4f35-a3f3-5b39e3186ff5-kube-api-access-vpn2f\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138656 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92c2898e-8e5b-4d26-b90f-97b357f5de80-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138683 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-config-volume\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138703 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/db5b0254-7015-4320-b782-ac764446265f-srv-cert\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138743 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-service-ca-bundle\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138753 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e5e62f0-77fd-446f-8166-154f715cea66-proxy-tls\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138811 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxmvv\" (UniqueName: \"kubernetes.io/projected/92c2898e-8e5b-4d26-b90f-97b357f5de80-kube-api-access-gxmvv\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138833 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7248143-1eb1-4858-a4a3-63d74fa4a94b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138853 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7248143-1eb1-4858-a4a3-63d74fa4a94b-config\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138869 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d845\" (UniqueName: \"kubernetes.io/projected/cc497af5-6bad-4dc2-a1a0-e574992346a5-kube-api-access-5d845\") pod \"dns-operator-744455d44c-q42hh\" (UID: \"cc497af5-6bad-4dc2-a1a0-e574992346a5\") " pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138892 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5603525a-e561-4243-90ae-eb910aa81d29-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138910 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m255s\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-kube-api-access-m255s\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138931 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3870a07-6c0e-4df8-8d96-159600ab5bd1-serving-cert\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138951 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5603525a-e561-4243-90ae-eb910aa81d29-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138968 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8b50b274-12b2-4ed3-ad0a-81365aeff72c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.138984 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-default-certificate\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139006 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-trusted-ca\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139022 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0f1f4b07-67d3-4920-8984-410096b9fe3d-proxy-tls\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139040 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-registration-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139057 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9tkt\" (UniqueName: \"kubernetes.io/projected/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-kube-api-access-x9tkt\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139079 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74f0b1b5-cf28-4dd7-9701-b7361caa5442-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139097 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139111 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-csi-data-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139131 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0f1f4b07-67d3-4920-8984-410096b9fe3d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139152 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7248143-1eb1-4858-a4a3-63d74fa4a94b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139169 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f0b1b5-cf28-4dd7-9701-b7361caa5442-config\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139189 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchd9\" (UniqueName: \"kubernetes.io/projected/797dc8be-e15a-49f7-b669-fecb0809759c-kube-api-access-zchd9\") pod \"migrator-59844c95c7-59g9q\" (UID: \"797dc8be-e15a-49f7-b669-fecb0809759c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139204 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvf8s\" (UniqueName: \"kubernetes.io/projected/db5b0254-7015-4320-b782-ac764446265f-kube-api-access-tvf8s\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139220 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qb7v\" (UniqueName: \"kubernetes.io/projected/5e5e62f0-77fd-446f-8166-154f715cea66-kube-api-access-5qb7v\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139240 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chsm8\" (UniqueName: \"kubernetes.io/projected/f3870a07-6c0e-4df8-8d96-159600ab5bd1-kube-api-access-chsm8\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139255 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f269f51f-578d-49f0-8696-c2c69e8763fd-certs\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139274 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c2898e-8e5b-4d26-b90f-97b357f5de80-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139290 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sqsk\" (UniqueName: \"kubernetes.io/projected/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-kube-api-access-8sqsk\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139307 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3fe840a-b55a-459a-98eb-f3263afd9184-trusted-ca\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139326 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f0ef5080-e168-4e13-92d5-8a41f51610da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139342 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139384 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f0ef5080-e168-4e13-92d5-8a41f51610da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139399 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139419 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwwfg\" (UniqueName: \"kubernetes.io/projected/6b42da42-0ae3-43d1-be76-e400b3a124ac-kube-api-access-xwwfg\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139437 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d7975b7a-932a-41de-91aa-06a5b67833a2-tmpfs\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139453 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-mountpoint-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139480 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-metrics-tls\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139503 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139523 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/96abee8c-2ffa-4329-9f61-03e783f933b2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139539 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7975b7a-932a-41de-91aa-06a5b67833a2-webhook-cert\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139560 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7975b7a-932a-41de-91aa-06a5b67833a2-apiservice-cert\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139578 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgnmc\" (UniqueName: \"kubernetes.io/projected/c3fe840a-b55a-459a-98eb-f3263afd9184-kube-api-access-lgnmc\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139597 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkn9k\" (UniqueName: \"kubernetes.io/projected/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-kube-api-access-dkn9k\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.139615 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1621fdc1-9b5b-4022-940a-7b8119be5b9d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cx5k6\" (UID: \"1621fdc1-9b5b-4022-940a-7b8119be5b9d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.140357 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5603525a-e561-4243-90ae-eb910aa81d29-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.140523 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7248143-1eb1-4858-a4a3-63d74fa4a94b-config\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.140633 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8b50b274-12b2-4ed3-ad0a-81365aeff72c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.140883 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3fe840a-b55a-459a-98eb-f3263afd9184-trusted-ca\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.140891 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47b29\" (UniqueName: \"kubernetes.io/projected/0f1f4b07-67d3-4920-8984-410096b9fe3d-kube-api-access-47b29\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.140919 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-profile-collector-cert\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.142676 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.143496 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b42da42-0ae3-43d1-be76-e400b3a124ac-serving-cert\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.143529 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-socket-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.143570 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5e5e62f0-77fd-446f-8166-154f715cea66-images\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.143593 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.143612 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5603525a-e561-4243-90ae-eb910aa81d29-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.143629 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/db5b0254-7015-4320-b782-ac764446265f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.145820 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.146254 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-config-volume\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.146354 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f0b1b5-cf28-4dd7-9701-b7361caa5442-config\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.146692 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-trusted-ca\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.146872 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/59925674-795d-4e20-b71b-351cf566550d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4lbsj\" (UID: \"59925674-795d-4e20-b71b-351cf566550d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.147055 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f0ef5080-e168-4e13-92d5-8a41f51610da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.147141 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.147189 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c2898e-8e5b-4d26-b90f-97b357f5de80-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.147647 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5e5e62f0-77fd-446f-8166-154f715cea66-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.148962 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.648944614 +0000 UTC m=+147.236922880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.150482 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d7975b7a-932a-41de-91aa-06a5b67833a2-tmpfs\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.152175 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5e5e62f0-77fd-446f-8166-154f715cea66-images\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.153235 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7248143-1eb1-4858-a4a3-63d74fa4a94b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.153325 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-certificates\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.153323 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1621fdc1-9b5b-4022-940a-7b8119be5b9d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cx5k6\" (UID: \"1621fdc1-9b5b-4022-940a-7b8119be5b9d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.153631 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5e5e62f0-77fd-446f-8166-154f715cea66-proxy-tls\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.154508 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/96abee8c-2ffa-4329-9f61-03e783f933b2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.159258 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.160410 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74f0b1b5-cf28-4dd7-9701-b7361caa5442-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.160550 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.161116 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/db5b0254-7015-4320-b782-ac764446265f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.161311 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-tls\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.161558 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f893fe9-73c9-4758-a561-617a0afee23c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fwxdh\" (UID: \"9f893fe9-73c9-4758-a561-617a0afee23c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.161643 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.162009 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-stats-auth\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.162290 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96abee8c-2ffa-4329-9f61-03e783f933b2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.162456 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3fe840a-b55a-459a-98eb-f3263afd9184-metrics-tls\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.162514 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cc497af5-6bad-4dc2-a1a0-e574992346a5-metrics-tls\") pod \"dns-operator-744455d44c-q42hh\" (UID: \"cc497af5-6bad-4dc2-a1a0-e574992346a5\") " pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.162609 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.162775 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.163201 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92c2898e-8e5b-4d26-b90f-97b357f5de80-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.163381 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7975b7a-932a-41de-91aa-06a5b67833a2-apiservice-cert\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.163823 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b42da42-0ae3-43d1-be76-e400b3a124ac-config\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.163862 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7975b7a-932a-41de-91aa-06a5b67833a2-webhook-cert\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.164127 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-default-certificate\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.164754 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.165305 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/db5b0254-7015-4320-b782-ac764446265f-srv-cert\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.165443 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-secret-volume\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.166529 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-metrics-certs\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.167058 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0f1f4b07-67d3-4920-8984-410096b9fe3d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.167183 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f0ef5080-e168-4e13-92d5-8a41f51610da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.167445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b50b274-12b2-4ed3-ad0a-81365aeff72c-serving-cert\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.167445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b42da42-0ae3-43d1-be76-e400b3a124ac-serving-cert\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.169316 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0f1f4b07-67d3-4920-8984-410096b9fe3d-proxy-tls\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.174327 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/55dbc750-3c63-44f9-ada1-aa282adb3562-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-kvssf\" (UID: \"55dbc750-3c63-44f9-ada1-aa282adb3562\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.175923 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcn7d\" (UniqueName: \"kubernetes.io/projected/55dbc750-3c63-44f9-ada1-aa282adb3562-kube-api-access-bcn7d\") pod \"control-plane-machine-set-operator-78cbb6b69f-kvssf\" (UID: \"55dbc750-3c63-44f9-ada1-aa282adb3562\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.176693 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5603525a-e561-4243-90ae-eb910aa81d29-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.193188 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sg7l\" (UniqueName: \"kubernetes.io/projected/8b50b274-12b2-4ed3-ad0a-81365aeff72c-kube-api-access-6sg7l\") pod \"openshift-config-operator-7777fb866f-s4gdr\" (UID: \"8b50b274-12b2-4ed3-ad0a-81365aeff72c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.220546 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.237943 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz9qp\" (UniqueName: \"kubernetes.io/projected/59925674-795d-4e20-b71b-351cf566550d-kube-api-access-tz9qp\") pod \"multus-admission-controller-857f4d67dd-4lbsj\" (UID: \"59925674-795d-4e20-b71b-351cf566550d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.244858 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245076 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-plugins-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245098 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/87433151-5993-4f35-a3f3-5b39e3186ff5-signing-key\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245119 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-srv-cert\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245160 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl5jl\" (UniqueName: \"kubernetes.io/projected/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-kube-api-access-cl5jl\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245177 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3870a07-6c0e-4df8-8d96-159600ab5bd1-config\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245194 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbtgm\" (UniqueName: \"kubernetes.io/projected/f269f51f-578d-49f0-8696-c2c69e8763fd-kube-api-access-kbtgm\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245215 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/87433151-5993-4f35-a3f3-5b39e3186ff5-signing-cabundle\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245234 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9znlm\" (UniqueName: \"kubernetes.io/projected/2e66dca3-d714-48b7-aceb-b9a826efcfdc-kube-api-access-9znlm\") pod \"ingress-canary-z9v9l\" (UID: \"2e66dca3-d714-48b7-aceb-b9a826efcfdc\") " pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245257 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-config-volume\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245279 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f269f51f-578d-49f0-8696-c2c69e8763fd-node-bootstrap-token\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245293 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpn2f\" (UniqueName: \"kubernetes.io/projected/87433151-5993-4f35-a3f3-5b39e3186ff5-kube-api-access-vpn2f\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245332 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3870a07-6c0e-4df8-8d96-159600ab5bd1-serving-cert\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245348 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-registration-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245366 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9tkt\" (UniqueName: \"kubernetes.io/projected/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-kube-api-access-x9tkt\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245385 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-csi-data-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245427 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chsm8\" (UniqueName: \"kubernetes.io/projected/f3870a07-6c0e-4df8-8d96-159600ab5bd1-kube-api-access-chsm8\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245443 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f269f51f-578d-49f0-8696-c2c69e8763fd-certs\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245474 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-mountpoint-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245496 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-metrics-tls\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245526 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-profile-collector-cert\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245542 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-socket-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245574 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq6cr\" (UniqueName: \"kubernetes.io/projected/7bc1102b-0f53-4b51-b5c4-eeb118249dea-kube-api-access-kq6cr\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.245588 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e66dca3-d714-48b7-aceb-b9a826efcfdc-cert\") pod \"ingress-canary-z9v9l\" (UID: \"2e66dca3-d714-48b7-aceb-b9a826efcfdc\") " pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.246249 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.746221481 +0000 UTC m=+147.334199747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.246543 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-plugins-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.247399 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-mountpoint-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.247453 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-csi-data-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.247487 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-registration-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.247558 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1102b-0f53-4b51-b5c4-eeb118249dea-socket-dir\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.247979 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/87433151-5993-4f35-a3f3-5b39e3186ff5-signing-cabundle\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.248224 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3870a07-6c0e-4df8-8d96-159600ab5bd1-config\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.248285 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-config-volume\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.250008 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-srv-cert\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.250216 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/87433151-5993-4f35-a3f3-5b39e3186ff5-signing-key\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.250374 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3870a07-6c0e-4df8-8d96-159600ab5bd1-serving-cert\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.250682 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e66dca3-d714-48b7-aceb-b9a826efcfdc-cert\") pod \"ingress-canary-z9v9l\" (UID: \"2e66dca3-d714-48b7-aceb-b9a826efcfdc\") " pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.250910 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-metrics-tls\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.251103 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f269f51f-578d-49f0-8696-c2c69e8763fd-node-bootstrap-token\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.251566 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f269f51f-578d-49f0-8696-c2c69e8763fd-certs\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.255685 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-profile-collector-cert\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.265639 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.272994 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.273473 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/96abee8c-2ffa-4329-9f61-03e783f933b2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.287300 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5kzwm"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.294194 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdsbt\" (UniqueName: \"kubernetes.io/projected/1621fdc1-9b5b-4022-940a-7b8119be5b9d-kube-api-access-xdsbt\") pod \"package-server-manager-789f6589d5-cx5k6\" (UID: \"1621fdc1-9b5b-4022-940a-7b8119be5b9d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.295469 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8whv\" (UniqueName: \"kubernetes.io/projected/9f893fe9-73c9-4758-a561-617a0afee23c-kube-api-access-q8whv\") pod \"cluster-samples-operator-665b6dd947-fwxdh\" (UID: \"9f893fe9-73c9-4758-a561-617a0afee23c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.295802 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxmvv\" (UniqueName: \"kubernetes.io/projected/92c2898e-8e5b-4d26-b90f-97b357f5de80-kube-api-access-gxmvv\") pod \"openshift-controller-manager-operator-756b6f6bc6-stbnh\" (UID: \"92c2898e-8e5b-4d26-b90f-97b357f5de80\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.312441 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74f0b1b5-cf28-4dd7-9701-b7361caa5442-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hhbql\" (UID: \"74f0b1b5-cf28-4dd7-9701-b7361caa5442\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.331683 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mxqrr"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.337394 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2fqh\" (UniqueName: \"kubernetes.io/projected/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-kube-api-access-n2fqh\") pod \"marketplace-operator-79b997595-l6fvp\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.345852 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9xjrk"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.347591 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.351704 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.851683019 +0000 UTC m=+147.439661285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.351763 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b5b1a61_6e93_4e2c_b317_ac90df8381aa.slice/crio-7f1fd90bec9b332855789779bd26dc19e95bb3533237a52c79ca4838589cd4a0 WatchSource:0}: Error finding container 7f1fd90bec9b332855789779bd26dc19e95bb3533237a52c79ca4838589cd4a0: Status 404 returned error can't find the container with id 7f1fd90bec9b332855789779bd26dc19e95bb3533237a52c79ca4838589cd4a0 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.353006 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pk7bd"] Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.354550 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d7e9659_3a05_461d_ab42_ceda65eb40e9.slice/crio-6b066455e48ca3fe45114ae75afec09264630928cbdd64ed5f474df82d340094 WatchSource:0}: Error finding container 6b066455e48ca3fe45114ae75afec09264630928cbdd64ed5f474df82d340094: Status 404 returned error can't find the container with id 6b066455e48ca3fe45114ae75afec09264630928cbdd64ed5f474df82d340094 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.360871 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdmmj\" (UniqueName: \"kubernetes.io/projected/27c80c4c-d3e4-4cce-a3ab-30d8073fdc42-kube-api-access-pdmmj\") pod \"kube-storage-version-migrator-operator-b67b599dd-vlm4v\" (UID: \"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.371931 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.373774 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wm9p\" (UniqueName: \"kubernetes.io/projected/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-kube-api-access-6wm9p\") pod \"collect-profiles-29405670-krkpx\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.377278 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75f8d46c_6fda_4058_95bb_da00aadb176b.slice/crio-e22923fa02f875dc731d2ba6983e57f55b319100f3d9fe982a68ef750886ebf0 WatchSource:0}: Error finding container e22923fa02f875dc731d2ba6983e57f55b319100f3d9fe982a68ef750886ebf0: Status 404 returned error can't find the container with id e22923fa02f875dc731d2ba6983e57f55b319100f3d9fe982a68ef750886ebf0 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.387469 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.409400 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.412738 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5603525a-e561-4243-90ae-eb910aa81d29-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5bqzh\" (UID: \"5603525a-e561-4243-90ae-eb910aa81d29\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.413770 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kzstg"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.431696 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.438006 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m255s\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-kube-api-access-m255s\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.448755 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.449362 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:04.949341585 +0000 UTC m=+147.537319851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.452259 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchd9\" (UniqueName: \"kubernetes.io/projected/797dc8be-e15a-49f7-b669-fecb0809759c-kube-api-access-zchd9\") pod \"migrator-59844c95c7-59g9q\" (UID: \"797dc8be-e15a-49f7-b669-fecb0809759c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.467797 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4552569_578c_4694_8e3b_1a3ec68380e9.slice/crio-7579eba1e499e0ab1681fe27ce12b818990e3be54f5f38dcf170230482a22719 WatchSource:0}: Error finding container 7579eba1e499e0ab1681fe27ce12b818990e3be54f5f38dcf170230482a22719: Status 404 returned error can't find the container with id 7579eba1e499e0ab1681fe27ce12b818990e3be54f5f38dcf170230482a22719 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.482151 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47b29\" (UniqueName: \"kubernetes.io/projected/0f1f4b07-67d3-4920-8984-410096b9fe3d-kube-api-access-47b29\") pod \"machine-config-controller-84d6567774-8c4fc\" (UID: \"0f1f4b07-67d3-4920-8984-410096b9fe3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.484008 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.513550 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.514464 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3fe840a-b55a-459a-98eb-f3263afd9184-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.526791 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.528883 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7248143-1eb1-4858-a4a3-63d74fa4a94b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j8cnv\" (UID: \"a7248143-1eb1-4858-a4a3-63d74fa4a94b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.529274 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.536039 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sqsh\" (UniqueName: \"kubernetes.io/projected/96abee8c-2ffa-4329-9f61-03e783f933b2-kube-api-access-8sqsh\") pod \"cluster-image-registry-operator-dc59b4c8b-jfp2b\" (UID: \"96abee8c-2ffa-4329-9f61-03e783f933b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.537137 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.539759 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qb7v\" (UniqueName: \"kubernetes.io/projected/5e5e62f0-77fd-446f-8166-154f715cea66-kube-api-access-5qb7v\") pod \"machine-config-operator-74547568cd-qp2rk\" (UID: \"5e5e62f0-77fd-446f-8166-154f715cea66\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.541693 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.543677 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.546609 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.550256 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.550815 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.05080048 +0000 UTC m=+147.638778746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.552250 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-v48xp"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.556786 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.559087 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvf8s\" (UniqueName: \"kubernetes.io/projected/db5b0254-7015-4320-b782-ac764446265f-kube-api-access-tvf8s\") pod \"olm-operator-6b444d44fb-ntw9s\" (UID: \"db5b0254-7015-4320-b782-ac764446265f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.560071 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jczpw"] Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.560454 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99d03802_d8df_4c6d_9788_e125e4b3dc84.slice/crio-c1dbe6d8f9361d02e9880713f821babbd35587d6ebb0974135182d500962e310 WatchSource:0}: Error finding container c1dbe6d8f9361d02e9880713f821babbd35587d6ebb0974135182d500962e310: Status 404 returned error can't find the container with id c1dbe6d8f9361d02e9880713f821babbd35587d6ebb0974135182d500962e310 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.563245 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" event={"ID":"a4552569-578c-4694-8e3b-1a3ec68380e9","Type":"ContainerStarted","Data":"7579eba1e499e0ab1681fe27ce12b818990e3be54f5f38dcf170230482a22719"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.565507 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9xjrk" event={"ID":"8d7e9659-3a05-461d-ab42-ceda65eb40e9","Type":"ContainerStarted","Data":"6b066455e48ca3fe45114ae75afec09264630928cbdd64ed5f474df82d340094"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.569780 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" event={"ID":"7b5b1a61-6e93-4e2c-b317-ac90df8381aa","Type":"ContainerStarted","Data":"7f1fd90bec9b332855789779bd26dc19e95bb3533237a52c79ca4838589cd4a0"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.573737 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" event={"ID":"9bad542b-4641-4db9-9a0e-bc1f80fd0a32","Type":"ContainerStarted","Data":"fc346c8fbfa27532c6815fff1748a3dc652838dd43a5127f48b6129c6179ebd5"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.575241 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" event={"ID":"e20cea1b-1075-4334-a1b6-3634ee774393","Type":"ContainerStarted","Data":"eba9479662a56c4be73ec7fc2ca0c9b1752df9657910e871cce4105d7e26dec1"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.582104 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4lbsj"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.583438 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5kzwm" event={"ID":"8e94a88b-c6a1-4edd-8759-136275441876","Type":"ContainerStarted","Data":"61807001d86420ddad66a73e9ff24d67363dfbf60aee033a9af3393b021ff340"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.591936 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" event={"ID":"67eed89a-6d2c-4b98-bf5b-13431b920d0c","Type":"ContainerStarted","Data":"83c4fce7ec7a9d45ead6a24fde7e69e67855735043befed228189f978d1fb64b"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.594642 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" event={"ID":"75f8d46c-6fda-4058-95bb-da00aadb176b","Type":"ContainerStarted","Data":"e22923fa02f875dc731d2ba6983e57f55b319100f3d9fe982a68ef750886ebf0"} Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.597116 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sqsk\" (UniqueName: \"kubernetes.io/projected/a3a550d3-7db4-456c-8f19-cf5e1cae5d2e-kube-api-access-8sqsk\") pod \"openshift-apiserver-operator-796bbdcf4f-7lmgq\" (UID: \"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.613445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjzts\" (UniqueName: \"kubernetes.io/projected/d7975b7a-932a-41de-91aa-06a5b67833a2-kube-api-access-xjzts\") pod \"packageserver-d55dfcdfc-pknqn\" (UID: \"d7975b7a-932a-41de-91aa-06a5b67833a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.636546 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgnmc\" (UniqueName: \"kubernetes.io/projected/c3fe840a-b55a-459a-98eb-f3263afd9184-kube-api-access-lgnmc\") pod \"ingress-operator-5b745b69d9-k6xw9\" (UID: \"c3fe840a-b55a-459a-98eb-f3263afd9184\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.649051 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf890cb5e_2cff_4904_b9f5_3ecd702e869e.slice/crio-d92ffcbb3226cbf663e02caaa8976ddf0c8638f88585bf793da6c91f8574ae81 WatchSource:0}: Error finding container d92ffcbb3226cbf663e02caaa8976ddf0c8638f88585bf793da6c91f8574ae81: Status 404 returned error can't find the container with id d92ffcbb3226cbf663e02caaa8976ddf0c8638f88585bf793da6c91f8574ae81 Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.650971 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.651128 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.651174 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.651205 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.651286 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.651545 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.151522755 +0000 UTC m=+147.739501021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.652570 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.654265 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.656097 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:32:04 crc kubenswrapper[4817]: W1128 14:32:04.657313 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59925674_795d_4e20_b71b_351cf566550d.slice/crio-c9b79128f2d5c0bf0f095b364fceb02bfd51a0b340042ad9dd58bceaf5365ddf WatchSource:0}: Error finding container c9b79128f2d5c0bf0f095b364fceb02bfd51a0b340042ad9dd58bceaf5365ddf: Status 404 returned error can't find the container with id c9b79128f2d5c0bf0f095b364fceb02bfd51a0b340042ad9dd58bceaf5365ddf Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.657359 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.657587 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.660942 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d845\" (UniqueName: \"kubernetes.io/projected/cc497af5-6bad-4dc2-a1a0-e574992346a5-kube-api-access-5d845\") pod \"dns-operator-744455d44c-q42hh\" (UID: \"cc497af5-6bad-4dc2-a1a0-e574992346a5\") " pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.661044 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.669788 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.672866 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkn9k\" (UniqueName: \"kubernetes.io/projected/b29cc89d-c829-41b8-bbb2-5a35b55ad6b4-kube-api-access-dkn9k\") pod \"router-default-5444994796-82mjn\" (UID: \"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4\") " pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.680283 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.700581 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.701332 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-bound-sa-token\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.707112 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.721607 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwwfg\" (UniqueName: \"kubernetes.io/projected/6b42da42-0ae3-43d1-be76-e400b3a124ac-kube-api-access-xwwfg\") pod \"authentication-operator-69f744f599-czfcx\" (UID: \"6b42da42-0ae3-43d1-be76-e400b3a124ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.739076 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpn2f\" (UniqueName: \"kubernetes.io/projected/87433151-5993-4f35-a3f3-5b39e3186ff5-kube-api-access-vpn2f\") pod \"service-ca-9c57cc56f-4jlk5\" (UID: \"87433151-5993-4f35-a3f3-5b39e3186ff5\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.754010 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.754441 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.254423097 +0000 UTC m=+147.842401363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.758206 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.760443 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbtgm\" (UniqueName: \"kubernetes.io/projected/f269f51f-578d-49f0-8696-c2c69e8763fd-kube-api-access-kbtgm\") pod \"machine-config-server-zkksb\" (UID: \"f269f51f-578d-49f0-8696-c2c69e8763fd\") " pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.773882 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.777400 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.783381 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chsm8\" (UniqueName: \"kubernetes.io/projected/f3870a07-6c0e-4df8-8d96-159600ab5bd1-kube-api-access-chsm8\") pod \"service-ca-operator-777779d784-s56j8\" (UID: \"f3870a07-6c0e-4df8-8d96-159600ab5bd1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.791982 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.799640 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.812407 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.815143 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl5jl\" (UniqueName: \"kubernetes.io/projected/229c92d3-dd9e-4b9e-8d87-f7fed988f8a6-kube-api-access-cl5jl\") pod \"dns-default-mbk86\" (UID: \"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6\") " pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.823660 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq6cr\" (UniqueName: \"kubernetes.io/projected/7bc1102b-0f53-4b51-b5c4-eeb118249dea-kube-api-access-kq6cr\") pod \"csi-hostpathplugin-mj798\" (UID: \"7bc1102b-0f53-4b51-b5c4-eeb118249dea\") " pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.848256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.851371 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.866191 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.866342 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.366322321 +0000 UTC m=+147.954300587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.866569 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.866882 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.366871585 +0000 UTC m=+147.954849851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.868517 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9znlm\" (UniqueName: \"kubernetes.io/projected/2e66dca3-d714-48b7-aceb-b9a826efcfdc-kube-api-access-9znlm\") pod \"ingress-canary-z9v9l\" (UID: \"2e66dca3-d714-48b7-aceb-b9a826efcfdc\") " pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.883032 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.886081 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9tkt\" (UniqueName: \"kubernetes.io/projected/98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b-kube-api-access-x9tkt\") pod \"catalog-operator-68c6474976-gdrnh\" (UID: \"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.888213 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.900015 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.910008 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.914564 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.923450 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zkksb" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.947099 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mj798" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.954921 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.955188 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-z9v9l" Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.967434 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:04 crc kubenswrapper[4817]: E1128 14:32:04.967869 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.467848867 +0000 UTC m=+148.055827133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.979736 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh"] Nov 28 14:32:04 crc kubenswrapper[4817]: I1128 14:32:04.984575 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.071239 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.071693 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.571678082 +0000 UTC m=+148.159656348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.176384 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.177142 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.67712167 +0000 UTC m=+148.265099946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.279308 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.279917 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.779899358 +0000 UTC m=+148.367877624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.314804 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.346597 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v"] Nov 28 14:32:05 crc kubenswrapper[4817]: W1128 14:32:05.361215 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f0b1b5_cf28_4dd7_9701_b7361caa5442.slice/crio-74d26a3f919c6409c4747b44044068320ca7be8854bba7854557615dc2f9c107 WatchSource:0}: Error finding container 74d26a3f919c6409c4747b44044068320ca7be8854bba7854557615dc2f9c107: Status 404 returned error can't find the container with id 74d26a3f919c6409c4747b44044068320ca7be8854bba7854557615dc2f9c107 Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.382031 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.382379 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.882358018 +0000 UTC m=+148.470336284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: W1128 14:32:05.428981 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27c80c4c_d3e4_4cce_a3ab_30d8073fdc42.slice/crio-6036675b5b79cd531332646cef834a97aa154f327c811fc5b222b3144a59777f WatchSource:0}: Error finding container 6036675b5b79cd531332646cef834a97aa154f327c811fc5b222b3144a59777f: Status 404 returned error can't find the container with id 6036675b5b79cd531332646cef834a97aa154f327c811fc5b222b3144a59777f Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.440411 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.458781 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.481293 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l6fvp"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.483908 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.484226 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:05.984212333 +0000 UTC m=+148.572190589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: W1128 14:32:05.495813 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod797dc8be_e15a_49f7_b669_fecb0809759c.slice/crio-2f7b76470474b979b79831197ade033426c18ee12a307fd740eafce6c24e0600 WatchSource:0}: Error finding container 2f7b76470474b979b79831197ade033426c18ee12a307fd740eafce6c24e0600: Status 404 returned error can't find the container with id 2f7b76470474b979b79831197ade033426c18ee12a307fd740eafce6c24e0600 Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.563618 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.584767 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.585027 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.08501165 +0000 UTC m=+148.672989916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.608428 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5kzwm" event={"ID":"8e94a88b-c6a1-4edd-8759-136275441876","Type":"ContainerStarted","Data":"061a50e1733fbe0ab8aa96d9ae39500097b56d589b0a5ed28e538646cbf81326"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.608479 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.615280 4817 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kzwm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.615313 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kzwm" podUID="8e94a88b-c6a1-4edd-8759-136275441876" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.617416 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.641711 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" event={"ID":"59925674-795d-4e20-b71b-351cf566550d","Type":"ContainerStarted","Data":"c9b79128f2d5c0bf0f095b364fceb02bfd51a0b340042ad9dd58bceaf5365ddf"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.643275 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" event={"ID":"99d03802-d8df-4c6d-9788-e125e4b3dc84","Type":"ContainerStarted","Data":"c1dbe6d8f9361d02e9880713f821babbd35587d6ebb0974135182d500962e310"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.649188 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" event={"ID":"7b5b1a61-6e93-4e2c-b317-ac90df8381aa","Type":"ContainerStarted","Data":"0835614862dd43cc8a8f16548e4e796f0ddb4e8f4436ad1f8eea57d88c9fdb14"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.652092 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" event={"ID":"74f0b1b5-cf28-4dd7-9701-b7361caa5442","Type":"ContainerStarted","Data":"74d26a3f919c6409c4747b44044068320ca7be8854bba7854557615dc2f9c107"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.657225 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" event={"ID":"e20cea1b-1075-4334-a1b6-3634ee774393","Type":"ContainerStarted","Data":"95c874cf6fa57f211c9c7d5a531cb92ca0d672b53fe962f46df3dba60626b17a"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.659299 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" event={"ID":"55dbc750-3c63-44f9-ada1-aa282adb3562","Type":"ContainerStarted","Data":"5229147cd091e68e859576e2d53e1d3eb83e6a2944e9d28a51bbb1d5b9a3136b"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.659340 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" event={"ID":"55dbc750-3c63-44f9-ada1-aa282adb3562","Type":"ContainerStarted","Data":"b3b0e604cbceef53d789d43f45f8f96959e077726e7abb9f333d1e128f908f5e"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.683449 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk"] Nov 28 14:32:05 crc kubenswrapper[4817]: W1128 14:32:05.686062 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26e26dfe_f9df_4fff_8eee_f8895b76e4a4.slice/crio-a3b9a8e2e253ab5889f9e369f7cfba5997c6ad4642decef32f71b1b97fb29b64 WatchSource:0}: Error finding container a3b9a8e2e253ab5889f9e369f7cfba5997c6ad4642decef32f71b1b97fb29b64: Status 404 returned error can't find the container with id a3b9a8e2e253ab5889f9e369f7cfba5997c6ad4642decef32f71b1b97fb29b64 Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.686665 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.688046 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.187997084 +0000 UTC m=+148.775975360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.688572 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" event={"ID":"f890cb5e-2cff-4904-b9f5-3ecd702e869e","Type":"ContainerStarted","Data":"d92ffcbb3226cbf663e02caaa8976ddf0c8638f88585bf793da6c91f8574ae81"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.690673 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" event={"ID":"75f8d46c-6fda-4058-95bb-da00aadb176b","Type":"ContainerStarted","Data":"443d887c8b338acad5107f12dd1248141e5c8718168a7e1383fd122ec860302a"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.691618 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.693056 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" event={"ID":"8b50b274-12b2-4ed3-ad0a-81365aeff72c","Type":"ContainerStarted","Data":"dd8276425b4a1bf934564d011e6463f7e5742238ccec183e2ea63f73097d10b1"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.693090 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" event={"ID":"8b50b274-12b2-4ed3-ad0a-81365aeff72c","Type":"ContainerStarted","Data":"561f6e6ae9d1edab43d533aad9048ff12697755a46e2d29d99c89a8ba22f2c81"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.694401 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" event={"ID":"e626351a-e0d2-48df-b470-e79a3f6f074e","Type":"ContainerStarted","Data":"bb2fb3975aaee95e364b78ae286a46add0127f7b53a524e63b129f19c53d7293"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.700456 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" event={"ID":"92c2898e-8e5b-4d26-b90f-97b357f5de80","Type":"ContainerStarted","Data":"752e590ad4a6359afd132fee96a43799ac5d4e997374510316ab9c8c85ec797f"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.702085 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-82mjn" event={"ID":"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4","Type":"ContainerStarted","Data":"b78f627a69024f6c6ec3d621fbc8e8372e3a40b9bed907c6784e13b5f003e29b"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.703072 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" event={"ID":"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42","Type":"ContainerStarted","Data":"6036675b5b79cd531332646cef834a97aa154f327c811fc5b222b3144a59777f"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.708515 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" event={"ID":"1621fdc1-9b5b-4022-940a-7b8119be5b9d","Type":"ContainerStarted","Data":"9a514ceca3b772174cfe7eda64fe249dd4e2e7bd916c99d0bdd3f93ae90b77f4"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.712450 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" event={"ID":"67eed89a-6d2c-4b98-bf5b-13431b920d0c","Type":"ContainerStarted","Data":"3910f73f0a2b063c292f32612e970b794daaef7400184c7652da212a453d687b"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.712785 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.725042 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9xjrk" event={"ID":"8d7e9659-3a05-461d-ab42-ceda65eb40e9","Type":"ContainerStarted","Data":"b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.727554 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" event={"ID":"9bad542b-4641-4db9-9a0e-bc1f80fd0a32","Type":"ContainerStarted","Data":"4c0fa8e123ba0eb2cc59533109cdd0048d33b95ace913e5cd1c1f4996814ce54"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.729150 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" event={"ID":"797dc8be-e15a-49f7-b669-fecb0809759c","Type":"ContainerStarted","Data":"2f7b76470474b979b79831197ade033426c18ee12a307fd740eafce6c24e0600"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.730405 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" event={"ID":"d52d05b7-f5b8-4fb7-bd09-e063e48c615d","Type":"ContainerStarted","Data":"4192e7b6cc583c2a1feeb60a76fe46232ad86eac6478fceeb3252aff789f1403"} Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.761862 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh"] Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.792295 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.793239 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.293210306 +0000 UTC m=+148.881188582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: I1128 14:32:05.894451 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:05 crc kubenswrapper[4817]: E1128 14:32:05.894854 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.394839124 +0000 UTC m=+148.982817390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:05 crc kubenswrapper[4817]: W1128 14:32:05.934345 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5603525a_e561_4243_90ae_eb910aa81d29.slice/crio-4ee184c5553e6cf67048e751105cb466655a3c7a28903ff6707d673b284ff108 WatchSource:0}: Error finding container 4ee184c5553e6cf67048e751105cb466655a3c7a28903ff6707d673b284ff108: Status 404 returned error can't find the container with id 4ee184c5553e6cf67048e751105cb466655a3c7a28903ff6707d673b284ff108 Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.007459 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.008023 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.508008361 +0000 UTC m=+149.095986617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.112212 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" podStartSLOduration=129.112194035 podStartE2EDuration="2m9.112194035s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.109341372 +0000 UTC m=+148.697319638" watchObservedRunningTime="2025-11-28 14:32:06.112194035 +0000 UTC m=+148.700172321" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.120073 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.120443 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.620428708 +0000 UTC m=+149.208406974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.176461 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" podStartSLOduration=129.176442211 podStartE2EDuration="2m9.176442211s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.159149425 +0000 UTC m=+148.747127691" watchObservedRunningTime="2025-11-28 14:32:06.176442211 +0000 UTC m=+148.764420477" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.176685 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4jlk5"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.189886 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5kzwm" podStartSLOduration=129.189871487 podStartE2EDuration="2m9.189871487s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.188245275 +0000 UTC m=+148.776223541" watchObservedRunningTime="2025-11-28 14:32:06.189871487 +0000 UTC m=+148.777849753" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.192165 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.229323 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.229426 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.729403556 +0000 UTC m=+149.317381832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.229631 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.230087 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.730077043 +0000 UTC m=+149.318055309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.331322 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.331517 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.831497986 +0000 UTC m=+149.419476252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.337538 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.338039 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.837905522 +0000 UTC m=+149.425883788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.352142 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-kvssf" podStartSLOduration=129.352115918 podStartE2EDuration="2m9.352115918s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.350674561 +0000 UTC m=+148.938652827" watchObservedRunningTime="2025-11-28 14:32:06.352115918 +0000 UTC m=+148.940094184" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.378940 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pk7bd" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.381473 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.382819 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q42hh"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.409229 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mbk86"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.416707 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.429039 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mxqrr" podStartSLOduration=129.42901806 podStartE2EDuration="2m9.42901806s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.403926703 +0000 UTC m=+148.991904969" watchObservedRunningTime="2025-11-28 14:32:06.42901806 +0000 UTC m=+149.016996326" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.429971 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.448240 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-s56j8"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.453895 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-czfcx"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.454352 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.454648 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:06.954630819 +0000 UTC m=+149.542609085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.484411 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.554054 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-9xjrk" podStartSLOduration=129.554038681 podStartE2EDuration="2m9.554038681s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.553454046 +0000 UTC m=+149.141432302" watchObservedRunningTime="2025-11-28 14:32:06.554038681 +0000 UTC m=+149.142016947" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.556320 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.556647 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.056634058 +0000 UTC m=+149.644612324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.560901 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.590177 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn"] Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.592295 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mj798"] Nov 28 14:32:06 crc kubenswrapper[4817]: W1128 14:32:06.604986 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3a550d3_7db4_456c_8f19_cf5e1cae5d2e.slice/crio-232a4a42f2d573673eda89f5e59783628f8aedffbee7ab8bfc355e105125cf5a WatchSource:0}: Error finding container 232a4a42f2d573673eda89f5e59783628f8aedffbee7ab8bfc355e105125cf5a: Status 404 returned error can't find the container with id 232a4a42f2d573673eda89f5e59783628f8aedffbee7ab8bfc355e105125cf5a Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.614507 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-z9v9l"] Nov 28 14:32:06 crc kubenswrapper[4817]: W1128 14:32:06.619730 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3fe840a_b55a_459a_98eb_f3263afd9184.slice/crio-eeee2b6cdcde4cb92fc7dfdfd90b26a797b4320de2a120da0d85d28acfc11149 WatchSource:0}: Error finding container eeee2b6cdcde4cb92fc7dfdfd90b26a797b4320de2a120da0d85d28acfc11149: Status 404 returned error can't find the container with id eeee2b6cdcde4cb92fc7dfdfd90b26a797b4320de2a120da0d85d28acfc11149 Nov 28 14:32:06 crc kubenswrapper[4817]: W1128 14:32:06.622184 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bc1102b_0f53_4b51_b5c4_eeb118249dea.slice/crio-54d15192eb83cf155e248b89269b19765c10e1b9cc814ea3706a5ca60797f3d0 WatchSource:0}: Error finding container 54d15192eb83cf155e248b89269b19765c10e1b9cc814ea3706a5ca60797f3d0: Status 404 returned error can't find the container with id 54d15192eb83cf155e248b89269b19765c10e1b9cc814ea3706a5ca60797f3d0 Nov 28 14:32:06 crc kubenswrapper[4817]: W1128 14:32:06.642852 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7975b7a_932a_41de_91aa_06a5b67833a2.slice/crio-2560216f50624e45509d925102f81bc569cfab5b09d520e36aa65b9c0d0683a8 WatchSource:0}: Error finding container 2560216f50624e45509d925102f81bc569cfab5b09d520e36aa65b9c0d0683a8: Status 404 returned error can't find the container with id 2560216f50624e45509d925102f81bc569cfab5b09d520e36aa65b9c0d0683a8 Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.656959 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.657333 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.157313192 +0000 UTC m=+149.745291458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.738775 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" event={"ID":"d52d05b7-f5b8-4fb7-bd09-e063e48c615d","Type":"ContainerStarted","Data":"0b9c1851322fe0c25a724c89367a391dd434a84f03b8746661f38a9d155720b2"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.739657 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" event={"ID":"f3870a07-6c0e-4df8-8d96-159600ab5bd1","Type":"ContainerStarted","Data":"69d69cd6527472deda30c70ee9931f265dedf981db4a69b5724960369ed8f40b"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.740669 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" event={"ID":"87433151-5993-4f35-a3f3-5b39e3186ff5","Type":"ContainerStarted","Data":"ebcdeb37013e2e2e68f1394bb6661d9d967caca474235ab8b7e52845a6ce976d"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.741261 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" event={"ID":"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e","Type":"ContainerStarted","Data":"232a4a42f2d573673eda89f5e59783628f8aedffbee7ab8bfc355e105125cf5a"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.741804 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" event={"ID":"d7975b7a-932a-41de-91aa-06a5b67833a2","Type":"ContainerStarted","Data":"2560216f50624e45509d925102f81bc569cfab5b09d520e36aa65b9c0d0683a8"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.742533 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" event={"ID":"a7248143-1eb1-4858-a4a3-63d74fa4a94b","Type":"ContainerStarted","Data":"ccb33b3a7384ce632e618c326d0cc314181edf012c1f0434c74288567fbc4a2d"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.759152 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.759800 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.259783972 +0000 UTC m=+149.847762238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.765559 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" event={"ID":"27c80c4c-d3e4-4cce-a3ab-30d8073fdc42","Type":"ContainerStarted","Data":"c76a8d3cae9c846034cc0decc6278ad1cddfa5e38981966b22c07a85b27d7f96"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.769662 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" event={"ID":"cc497af5-6bad-4dc2-a1a0-e574992346a5","Type":"ContainerStarted","Data":"2eacaac2586fb803f24ab617c3679a2315cd9e7502373ec3348cf6488adc30ab"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.776239 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8fe170619b9340db40854b165407f9e804400254ff00d42573fc02ba2e836610"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.783473 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" event={"ID":"db5b0254-7015-4320-b782-ac764446265f","Type":"ContainerStarted","Data":"8d25dc62a9fb1a73be1290d4e5f94ee206747ac689aeb5e16290def7ff22fc83"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.797915 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" event={"ID":"92c2898e-8e5b-4d26-b90f-97b357f5de80","Type":"ContainerStarted","Data":"349aca827beeccf8445ffd2f0f576a27823673f5cdc3edac92136988a03e56f9"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.813389 4817 generic.go:334] "Generic (PLEG): container finished" podID="99d03802-d8df-4c6d-9788-e125e4b3dc84" containerID="f15db928eba691bdfb65bc06d2bc595fb113e0d04b7b75a89c8b2dfbd13bb04a" exitCode=0 Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.813454 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" event={"ID":"99d03802-d8df-4c6d-9788-e125e4b3dc84","Type":"ContainerDied","Data":"f15db928eba691bdfb65bc06d2bc595fb113e0d04b7b75a89c8b2dfbd13bb04a"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.830105 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" event={"ID":"e20cea1b-1075-4334-a1b6-3634ee774393","Type":"ContainerStarted","Data":"1da308d0e83bf989602e8f4a6044f50b166bf3eba71cfb793b11be5617e833be"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.834838 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" event={"ID":"6b42da42-0ae3-43d1-be76-e400b3a124ac","Type":"ContainerStarted","Data":"0a474d57a482154c30102532f12a578e0af2e6b7ec00be7b8ca58d62fb072251"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.838230 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" event={"ID":"a4552569-578c-4694-8e3b-1a3ec68380e9","Type":"ContainerStarted","Data":"1d068147cbbf0344fc9f546245c4f2fbc1bcc6db806a51b5fa0f7721ca0d40a9"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.839181 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.840376 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" event={"ID":"96abee8c-2ffa-4329-9f61-03e783f933b2","Type":"ContainerStarted","Data":"ec8c6c5e4d7386d2f066d4d29671df48e1d2a3c0be2053d2dded5fbbab4cef50"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.840400 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" event={"ID":"96abee8c-2ffa-4329-9f61-03e783f933b2","Type":"ContainerStarted","Data":"88821d4ceb902b14ad19b62ec6f4461b1cb484f1680d1fa19ca88736fa81ef8d"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.841535 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-z9v9l" event={"ID":"2e66dca3-d714-48b7-aceb-b9a826efcfdc","Type":"ContainerStarted","Data":"98e384ffa9de499297fd3a308e327b62a2e59d4bd847aaa66abf511aea75674a"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.842963 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" event={"ID":"5e5e62f0-77fd-446f-8166-154f715cea66","Type":"ContainerStarted","Data":"c967515c53e388157b88f0e4c5e6485ccb23cd3ac076580c3f67d7b7e7c93639"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.842999 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" event={"ID":"5e5e62f0-77fd-446f-8166-154f715cea66","Type":"ContainerStarted","Data":"06221377bc88d0f3b44cd81159afdfaa9793a4800bd1f5fd5d6c1e915db6b528"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.844172 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" event={"ID":"9f893fe9-73c9-4758-a561-617a0afee23c","Type":"ContainerStarted","Data":"398793789e2c30d6ce646b5224f63182055069d379b8f79e7a470f173d138928"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.846781 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" event={"ID":"59925674-795d-4e20-b71b-351cf566550d","Type":"ContainerStarted","Data":"0826aa740ec3b6565dd3736b52bd68ae1e9f98521d1b895b2750de345d770592"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.849100 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" event={"ID":"f890cb5e-2cff-4904-b9f5-3ecd702e869e","Type":"ContainerStarted","Data":"b40ad03709c2ac77d4278d1720ae5d7537ea4c1692cdb9aafc31af60f08a9802"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.849891 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.853691 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.857523 4817 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jczpw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.857577 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" podUID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.860853 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.861509 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.361493733 +0000 UTC m=+149.949471999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.863839 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" event={"ID":"0f1f4b07-67d3-4920-8984-410096b9fe3d","Type":"ContainerStarted","Data":"50edaf213d4e7d6fb7ab19f50d0875cb716f8fbba3168cb1ebbb2989a679d90f"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.863876 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" event={"ID":"0f1f4b07-67d3-4920-8984-410096b9fe3d","Type":"ContainerStarted","Data":"090fe08d1656458132a78a0387c22ac1dec630c1208f21ed6ca0ef1452649c26"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.869301 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zkksb" event={"ID":"f269f51f-578d-49f0-8696-c2c69e8763fd","Type":"ContainerStarted","Data":"9449bc5564471a842cacf2252ebac9f873a4868e4e027d984a3bf8098b2fc199"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.869343 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zkksb" event={"ID":"f269f51f-578d-49f0-8696-c2c69e8763fd","Type":"ContainerStarted","Data":"252cea929563e6b679990ffa883579044f954fd07e1e3d31c14820537dde18f5"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.873827 4817 generic.go:334] "Generic (PLEG): container finished" podID="8b50b274-12b2-4ed3-ad0a-81365aeff72c" containerID="dd8276425b4a1bf934564d011e6463f7e5742238ccec183e2ea63f73097d10b1" exitCode=0 Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.873887 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" event={"ID":"8b50b274-12b2-4ed3-ad0a-81365aeff72c","Type":"ContainerDied","Data":"dd8276425b4a1bf934564d011e6463f7e5742238ccec183e2ea63f73097d10b1"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.878441 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1d95190888fd1f2277805600feab50b3cab982b40648ac2daf0fa9ef7f6c979f"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.891613 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.926715 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" event={"ID":"797dc8be-e15a-49f7-b669-fecb0809759c","Type":"ContainerStarted","Data":"9c3c8689b15cac54afffe69cc3d50e4677433d6a439581bd699100222aec949f"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.927363 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" podStartSLOduration=129.9273458 podStartE2EDuration="2m9.9273458s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.923987273 +0000 UTC m=+149.511965539" watchObservedRunningTime="2025-11-28 14:32:06.9273458 +0000 UTC m=+149.515324066" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.927517 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" podStartSLOduration=130.927512454 podStartE2EDuration="2m10.927512454s" podCreationTimestamp="2025-11-28 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.891396654 +0000 UTC m=+149.479374920" watchObservedRunningTime="2025-11-28 14:32:06.927512454 +0000 UTC m=+149.515490720" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.950980 4817 generic.go:334] "Generic (PLEG): container finished" podID="e626351a-e0d2-48df-b470-e79a3f6f074e" containerID="51359f0e1374f2ae08f03490546ecf5315e0cdb66d524f84f5fa0b9485b5d28f" exitCode=0 Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.951755 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" event={"ID":"e626351a-e0d2-48df-b470-e79a3f6f074e","Type":"ContainerDied","Data":"51359f0e1374f2ae08f03490546ecf5315e0cdb66d524f84f5fa0b9485b5d28f"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.958406 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" event={"ID":"5603525a-e561-4243-90ae-eb910aa81d29","Type":"ContainerStarted","Data":"7c8e508077de2bf925f08b58b015091994023a3dac81162c489c117a45c22090"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.958457 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" event={"ID":"5603525a-e561-4243-90ae-eb910aa81d29","Type":"ContainerStarted","Data":"4ee184c5553e6cf67048e751105cb466655a3c7a28903ff6707d673b284ff108"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.962628 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b588079c966dd7488c1a307bd130846c5b878066b79dace61e279e73604a984a"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.964273 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:06 crc kubenswrapper[4817]: E1128 14:32:06.968816 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.468802768 +0000 UTC m=+150.056781034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.969632 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" event={"ID":"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b","Type":"ContainerStarted","Data":"22650f8c742cdaf025ac9b7b98e536530b7cdc0fd46f94c1bec1b2199454cff2"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.974517 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mbk86" event={"ID":"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6","Type":"ContainerStarted","Data":"f7e2983da3b2f5645929183a7de10087683b93511c1010eca72fccd64b3504de"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.977734 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" event={"ID":"26e26dfe-f9df-4fff-8eee-f8895b76e4a4","Type":"ContainerStarted","Data":"a3b9a8e2e253ab5889f9e369f7cfba5997c6ad4642decef32f71b1b97fb29b64"} Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.982396 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.982673 4817 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-l6fvp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 28 14:32:06 crc kubenswrapper[4817]: I1128 14:32:06.982735 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.013232 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-82mjn" event={"ID":"b29cc89d-c829-41b8-bbb2-5a35b55ad6b4","Type":"ContainerStarted","Data":"7382a67436143154a780895da951932e800fc92c3f9e2313e56d4522a33e6c22"} Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.015451 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-skq7g" podStartSLOduration=130.01542818 podStartE2EDuration="2m10.01542818s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:06.947293694 +0000 UTC m=+149.535271960" watchObservedRunningTime="2025-11-28 14:32:07.01542818 +0000 UTC m=+149.603406446" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.032557 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mj798" event={"ID":"7bc1102b-0f53-4b51-b5c4-eeb118249dea","Type":"ContainerStarted","Data":"54d15192eb83cf155e248b89269b19765c10e1b9cc814ea3706a5ca60797f3d0"} Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.059768 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" event={"ID":"9bad542b-4641-4db9-9a0e-bc1f80fd0a32","Type":"ContainerStarted","Data":"0328e27b9768a0f12adaad192d07692682f92b76d0ce98dee8cd6a2e2cdcbf00"} Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.066679 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.067331 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.567314717 +0000 UTC m=+150.155292983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.073367 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" event={"ID":"c3fe840a-b55a-459a-98eb-f3263afd9184","Type":"ContainerStarted","Data":"eeee2b6cdcde4cb92fc7dfdfd90b26a797b4320de2a120da0d85d28acfc11149"} Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.079178 4817 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kzwm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.079230 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kzwm" podUID="8e94a88b-c6a1-4edd-8759-136275441876" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.092332 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jfp2b" podStartSLOduration=130.092315911 podStartE2EDuration="2m10.092315911s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.090135695 +0000 UTC m=+149.678113961" watchObservedRunningTime="2025-11-28 14:32:07.092315911 +0000 UTC m=+149.680294177" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.143429 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" podStartSLOduration=130.143403798 podStartE2EDuration="2m10.143403798s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.137840744 +0000 UTC m=+149.725819010" watchObservedRunningTime="2025-11-28 14:32:07.143403798 +0000 UTC m=+149.731382064" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.170966 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.175731 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.67570062 +0000 UTC m=+150.263678876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.272003 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5bqzh" podStartSLOduration=130.271987821 podStartE2EDuration="2m10.271987821s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.252503409 +0000 UTC m=+149.840481675" watchObservedRunningTime="2025-11-28 14:32:07.271987821 +0000 UTC m=+149.859966087" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.273046 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.273331 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.773320535 +0000 UTC m=+150.361298801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.361768 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ffg76" podStartSLOduration=131.361754394 podStartE2EDuration="2m11.361754394s" podCreationTimestamp="2025-11-28 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.356080668 +0000 UTC m=+149.944058934" watchObservedRunningTime="2025-11-28 14:32:07.361754394 +0000 UTC m=+149.949732660" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.380374 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.380662 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.880648701 +0000 UTC m=+150.468626967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.442598 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-82mjn" podStartSLOduration=130.442583287 podStartE2EDuration="2m10.442583287s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.441691014 +0000 UTC m=+150.029669280" watchObservedRunningTime="2025-11-28 14:32:07.442583287 +0000 UTC m=+150.030561553" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.462559 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" podStartSLOduration=130.462541051 podStartE2EDuration="2m10.462541051s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.393666557 +0000 UTC m=+149.981644823" watchObservedRunningTime="2025-11-28 14:32:07.462541051 +0000 UTC m=+150.050519317" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.489149 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.489493 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:07.989477616 +0000 UTC m=+150.577455882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.591323 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.591660 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.091649318 +0000 UTC m=+150.679627584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.692102 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.692464 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.192443876 +0000 UTC m=+150.780422142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.773365 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zkksb" podStartSLOduration=6.773347321 podStartE2EDuration="6.773347321s" podCreationTimestamp="2025-11-28 14:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:07.504545364 +0000 UTC m=+150.092523630" watchObservedRunningTime="2025-11-28 14:32:07.773347321 +0000 UTC m=+150.361325587" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.823836 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.824228 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.324218541 +0000 UTC m=+150.912196807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.884212 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.907411 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:07 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:07 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:07 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.907473 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:07 crc kubenswrapper[4817]: I1128 14:32:07.924589 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:07 crc kubenswrapper[4817]: E1128 14:32:07.925273 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.425238535 +0000 UTC m=+151.013216801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.027982 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.028732 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.528700421 +0000 UTC m=+151.116678687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.098092 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8b0a1df1669bae8793f7d559b8b89f4c57abc3167296712ab13d243e87104516"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.130678 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.131077 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.631060889 +0000 UTC m=+151.219039155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.146620 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4lbsj" event={"ID":"59925674-795d-4e20-b71b-351cf566550d","Type":"ContainerStarted","Data":"4ead062bd370b1b1204199db4c58552232add3d3e73c11ce8f9d76f6680b4854"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.156522 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" event={"ID":"6b42da42-0ae3-43d1-be76-e400b3a124ac","Type":"ContainerStarted","Data":"ad6eac0c33ce42c45f594a0dfc83957fee776019f38316611aa2a903b33ba916"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.182150 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" event={"ID":"74f0b1b5-cf28-4dd7-9701-b7361caa5442","Type":"ContainerStarted","Data":"b181d023b24ddb42df3cbc0b85173b7955fea1552045b6eb40ebf8bebc57d06c"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.235031 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.239213 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.739186455 +0000 UTC m=+151.327164711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.249069 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" event={"ID":"a7248143-1eb1-4858-a4a3-63d74fa4a94b","Type":"ContainerStarted","Data":"0135c86ad7466cf1357581a45304939ab22dd49af446bdad58509fa615864d82"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.270167 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" event={"ID":"cc497af5-6bad-4dc2-a1a0-e574992346a5","Type":"ContainerStarted","Data":"3eaeaa0a2b98d6fa1a6da3718e532ca057922119e75a9abca8fd38d81613d97b"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.326056 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" event={"ID":"c3fe840a-b55a-459a-98eb-f3263afd9184","Type":"ContainerStarted","Data":"1d429cf38345cc8af8aaec3611db003fcfe26010a46e6621b151e79c73c8987c"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.329324 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-czfcx" podStartSLOduration=132.329308857 podStartE2EDuration="2m12.329308857s" podCreationTimestamp="2025-11-28 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.326437223 +0000 UTC m=+150.914415489" watchObservedRunningTime="2025-11-28 14:32:08.329308857 +0000 UTC m=+150.917287123" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.337227 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.339403 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.839389517 +0000 UTC m=+151.427367783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.373991 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" event={"ID":"9f893fe9-73c9-4758-a561-617a0afee23c","Type":"ContainerStarted","Data":"6aa08eea511e7b54bde6045fee4cf86b02827e5e25559a5e46210eff79837d2e"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.382582 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hhbql" podStartSLOduration=131.382544599 podStartE2EDuration="2m11.382544599s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.36395664 +0000 UTC m=+150.951934906" watchObservedRunningTime="2025-11-28 14:32:08.382544599 +0000 UTC m=+150.970522855" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.396980 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" event={"ID":"db5b0254-7015-4320-b782-ac764446265f","Type":"ContainerStarted","Data":"200ef68b4b7deeb9079e12941344176cb0fecba8445cafecb00df648865ddb73"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.397944 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.405694 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j8cnv" podStartSLOduration=131.405677305 podStartE2EDuration="2m11.405677305s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.404831163 +0000 UTC m=+150.992809429" watchObservedRunningTime="2025-11-28 14:32:08.405677305 +0000 UTC m=+150.993655571" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.413344 4817 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ntw9s container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.413415 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" podUID="db5b0254-7015-4320-b782-ac764446265f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.448051 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.449270 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:08.949257928 +0000 UTC m=+151.537236184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.451998 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" event={"ID":"d7975b7a-932a-41de-91aa-06a5b67833a2","Type":"ContainerStarted","Data":"4d0a8023dccd5f0d3157e20dc0eda86e6275199f929d35270ce722be92523cb0"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.452968 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.471106 4817 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pknqn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.471174 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" podUID="d7975b7a-932a-41de-91aa-06a5b67833a2" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.479597 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" event={"ID":"26e26dfe-f9df-4fff-8eee-f8895b76e4a4","Type":"ContainerStarted","Data":"332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.480672 4817 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-l6fvp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.480700 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.481730 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"80a3b93750927dc06a251fbf0c2e0339f50585abc79ff49989fcceabbe9daba0"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.493679 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" podStartSLOduration=131.493661052 podStartE2EDuration="2m11.493661052s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.491977709 +0000 UTC m=+151.079955975" watchObservedRunningTime="2025-11-28 14:32:08.493661052 +0000 UTC m=+151.081639308" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.504978 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" event={"ID":"a3a550d3-7db4-456c-8f19-cf5e1cae5d2e","Type":"ContainerStarted","Data":"476abcf7cf40079ca51012a3d3dc7ee1115d6888d0551517135b8d31f1abddd1"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.577442 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" podStartSLOduration=131.577423621 podStartE2EDuration="2m11.577423621s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.576021935 +0000 UTC m=+151.164000201" watchObservedRunningTime="2025-11-28 14:32:08.577423621 +0000 UTC m=+151.165401887" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.581596 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.581963 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.081930597 +0000 UTC m=+151.669908863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.582686 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.584838 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" event={"ID":"f3870a07-6c0e-4df8-8d96-159600ab5bd1","Type":"ContainerStarted","Data":"806a7c470e0685cfe819bd54eb02154f0016d1ca3f85170b54f75532d3ee50a1"} Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.603215 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.103174434 +0000 UTC m=+151.691152700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.647343 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" event={"ID":"98c3d995-59e0-4ad7-a8b5-18dd6bc7f26b","Type":"ContainerStarted","Data":"e6e7bb09bc5fc3d24476db088752d0e38b436d7273932bc2eed06d9ee2339c75"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.648409 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.692306 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.693566 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.193545733 +0000 UTC m=+151.781523999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.697623 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" podStartSLOduration=131.697601928 podStartE2EDuration="2m11.697601928s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.686756128 +0000 UTC m=+151.274734394" watchObservedRunningTime="2025-11-28 14:32:08.697601928 +0000 UTC m=+151.285580204" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.701102 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" event={"ID":"1621fdc1-9b5b-4022-940a-7b8119be5b9d","Type":"ContainerStarted","Data":"9b7287f694da56f3b5a5670a08f5d84030b6c958708c9f57d1b0ee94b5bc744d"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.731173 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.740891 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2bd4f3bd261a5a1e2a31d47817d3096395400c3e2addc67717b82f554dc303cd"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.771803 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" event={"ID":"87433151-5993-4f35-a3f3-5b39e3186ff5","Type":"ContainerStarted","Data":"3d9f9df9729c75d002fb103279c96fb6f062dec7a6b6e255f65cc3d6e39bf7d5"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.788950 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-z9v9l" event={"ID":"2e66dca3-d714-48b7-aceb-b9a826efcfdc","Type":"ContainerStarted","Data":"2a3447f8c6110c284ed5e40f7872bb0331d4b5ca17fcfa01ebd73a341f6047e1"} Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.789384 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" podStartSLOduration=131.789357712 podStartE2EDuration="2m11.789357712s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.761518445 +0000 UTC m=+151.349496711" watchObservedRunningTime="2025-11-28 14:32:08.789357712 +0000 UTC m=+151.377335978" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.795536 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.814997 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.314979292 +0000 UTC m=+151.902957568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.829023 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7lmgq" podStartSLOduration=132.829000144 podStartE2EDuration="2m12.829000144s" podCreationTimestamp="2025-11-28 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.827626078 +0000 UTC m=+151.415604344" watchObservedRunningTime="2025-11-28 14:32:08.829000144 +0000 UTC m=+151.416978410" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.861483 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vlm4v" podStartSLOduration=131.86146239 podStartE2EDuration="2m11.86146239s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.860152976 +0000 UTC m=+151.448131252" watchObservedRunningTime="2025-11-28 14:32:08.86146239 +0000 UTC m=+151.449440656" Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.879053 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd52d05b7_f5b8_4fb7_bd09_e063e48c615d.slice/crio-conmon-0b9c1851322fe0c25a724c89367a391dd434a84f03b8746661f38a9d155720b2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd52d05b7_f5b8_4fb7_bd09_e063e48c615d.slice/crio-0b9c1851322fe0c25a724c89367a391dd434a84f03b8746661f38a9d155720b2.scope\": RecentStats: unable to find data in memory cache]" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.896034 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:08 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:08 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:08 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.896080 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.897204 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:08 crc kubenswrapper[4817]: E1128 14:32:08.899662 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.399643634 +0000 UTC m=+151.987621900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:08 crc kubenswrapper[4817]: I1128 14:32:08.957608 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-4jlk5" podStartSLOduration=131.957590067 podStartE2EDuration="2m11.957590067s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.927287566 +0000 UTC m=+151.515265832" watchObservedRunningTime="2025-11-28 14:32:08.957590067 +0000 UTC m=+151.545568333" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.000538 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.000962 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.500949595 +0000 UTC m=+152.088927871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.001902 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" podStartSLOduration=129.001875878 podStartE2EDuration="2m9.001875878s" podCreationTimestamp="2025-11-28 14:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.000317188 +0000 UTC m=+151.588295474" watchObservedRunningTime="2025-11-28 14:32:09.001875878 +0000 UTC m=+151.589854144" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.003151 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gdrnh" podStartSLOduration=132.003143231 podStartE2EDuration="2m12.003143231s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:08.958669695 +0000 UTC m=+151.546647961" watchObservedRunningTime="2025-11-28 14:32:09.003143231 +0000 UTC m=+151.591121497" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.070396 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-stbnh" podStartSLOduration=132.070381924 podStartE2EDuration="2m12.070381924s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.069785088 +0000 UTC m=+151.657763354" watchObservedRunningTime="2025-11-28 14:32:09.070381924 +0000 UTC m=+151.658360190" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.108516 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.109167 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.609145253 +0000 UTC m=+152.197123519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.113527 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-z9v9l" podStartSLOduration=8.113512095 podStartE2EDuration="8.113512095s" podCreationTimestamp="2025-11-28 14:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.113077374 +0000 UTC m=+151.701055640" watchObservedRunningTime="2025-11-28 14:32:09.113512095 +0000 UTC m=+151.701490361" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.137938 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-s56j8" podStartSLOduration=132.137921424 podStartE2EDuration="2m12.137921424s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.137070102 +0000 UTC m=+151.725048358" watchObservedRunningTime="2025-11-28 14:32:09.137921424 +0000 UTC m=+151.725899690" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.213215 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.213586 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.713569824 +0000 UTC m=+152.301548090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.314347 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.314751 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.81473315 +0000 UTC m=+152.402711416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.428513 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.429107 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:09.929091307 +0000 UTC m=+152.517069573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.530278 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.530470 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.030452159 +0000 UTC m=+152.618430425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.530547 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.530849 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.030843069 +0000 UTC m=+152.618821335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.631533 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.631700 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.131677408 +0000 UTC m=+152.719655684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.631832 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.632074 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.132066618 +0000 UTC m=+152.720044884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.732369 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.732585 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.732871 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.232818904 +0000 UTC m=+152.820797180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.799947 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mj798" event={"ID":"7bc1102b-0f53-4b51-b5c4-eeb118249dea","Type":"ContainerStarted","Data":"237dd075744e5f289f10c2ee4e080a3a4992bfeecc378badcf977ae193700a45"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.801790 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" event={"ID":"c3fe840a-b55a-459a-98eb-f3263afd9184","Type":"ContainerStarted","Data":"397f461cbfec4295c2c90d6e9c0e6f0c2bc7389be5272a18a140621abe2c210b"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.806912 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" event={"ID":"5e5e62f0-77fd-446f-8166-154f715cea66","Type":"ContainerStarted","Data":"f6cc98c18e6836cde537843d1c8e226d68e8b8ade0b24da4fdf23e0862613e51"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.824701 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" event={"ID":"e626351a-e0d2-48df-b470-e79a3f6f074e","Type":"ContainerStarted","Data":"13bc41de2597563dad187a7055591c111e7dc4c7877dd8b28bb45f9d27010d38"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.824777 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" event={"ID":"e626351a-e0d2-48df-b470-e79a3f6f074e","Type":"ContainerStarted","Data":"9b37d7c0b341764f5b9a8d3f3d5cfc2c78688ba8149ed0452d0090e7e4b99956"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.831903 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mbk86" event={"ID":"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6","Type":"ContainerStarted","Data":"e865ff05822cfe5ff1ac5e9f3c0f6e0390370d0679c4a16511ebae7b77ffd76c"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.831943 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mbk86" event={"ID":"229c92d3-dd9e-4b9e-8d87-f7fed988f8a6","Type":"ContainerStarted","Data":"40757ace76381acfd425e545b980171c4def08a9417306e02c1ee43ff564f97b"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.832634 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.835932 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.836791 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.336772533 +0000 UTC m=+152.924750799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.842545 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k6xw9" podStartSLOduration=132.842524271 podStartE2EDuration="2m12.842524271s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.841149566 +0000 UTC m=+152.429127832" watchObservedRunningTime="2025-11-28 14:32:09.842524271 +0000 UTC m=+152.430502537" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.845176 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" event={"ID":"8b50b274-12b2-4ed3-ad0a-81365aeff72c","Type":"ContainerStarted","Data":"ef03f210a0e7205e69d6f71aafa8bb61290860ef98040a230cd5b4491e53aaa1"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.845841 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.877176 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qp2rk" podStartSLOduration=132.877159714 podStartE2EDuration="2m12.877159714s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.875787768 +0000 UTC m=+152.463766034" watchObservedRunningTime="2025-11-28 14:32:09.877159714 +0000 UTC m=+152.465137980" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.878271 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" event={"ID":"1621fdc1-9b5b-4022-940a-7b8119be5b9d","Type":"ContainerStarted","Data":"15aea894952347f8f629d70cdb7afd779d15910372dca4165cfa0657320874d2"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.878801 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.893928 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:09 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:09 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:09 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.893981 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.913899 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fwxdh" event={"ID":"9f893fe9-73c9-4758-a561-617a0afee23c","Type":"ContainerStarted","Data":"b03712d8e929dbe5f51c4870cc334cf92db197ed29a69e7f43be8cccf51d6b34"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.936898 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:09 crc kubenswrapper[4817]: E1128 14:32:09.938347 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.43833202 +0000 UTC m=+153.026310286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.950139 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" event={"ID":"99d03802-d8df-4c6d-9788-e125e4b3dc84","Type":"ContainerStarted","Data":"6b2095d2992d6675946ba4c6edcc2c8ed26b6dcf22eaf6d9c4d4cdd5725a3ea3"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.964531 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" event={"ID":"cc497af5-6bad-4dc2-a1a0-e574992346a5","Type":"ContainerStarted","Data":"05278a62e20903f67de5f8fc6215c72b3da2b065de19b76d1661bb44eb505696"} Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.982048 4817 generic.go:334] "Generic (PLEG): container finished" podID="d52d05b7-f5b8-4fb7-bd09-e063e48c615d" containerID="0b9c1851322fe0c25a724c89367a391dd434a84f03b8746661f38a9d155720b2" exitCode=0 Nov 28 14:32:09 crc kubenswrapper[4817]: I1128 14:32:09.982112 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" event={"ID":"d52d05b7-f5b8-4fb7-bd09-e063e48c615d","Type":"ContainerDied","Data":"0b9c1851322fe0c25a724c89367a391dd434a84f03b8746661f38a9d155720b2"} Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.005187 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" event={"ID":"797dc8be-e15a-49f7-b669-fecb0809759c","Type":"ContainerStarted","Data":"7ea0a20d82d53dae6a41e3f174cadcb834acfe83d074bb795a240070e863531d"} Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.013692 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8c4fc" event={"ID":"0f1f4b07-67d3-4920-8984-410096b9fe3d","Type":"ContainerStarted","Data":"085d180fb59033eb61e70f783cc920ae48f2a282d76c5dfc00203c44119c9dd2"} Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.019889 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.042007 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.047675 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.547657657 +0000 UTC m=+153.135635923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.062912 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-mbk86" podStartSLOduration=9.06289539 podStartE2EDuration="9.06289539s" podCreationTimestamp="2025-11-28 14:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:09.997583657 +0000 UTC m=+152.585561923" watchObservedRunningTime="2025-11-28 14:32:10.06289539 +0000 UTC m=+152.650873656" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.063924 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" podStartSLOduration=134.063917336 podStartE2EDuration="2m14.063917336s" podCreationTimestamp="2025-11-28 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:10.06328035 +0000 UTC m=+152.651258616" watchObservedRunningTime="2025-11-28 14:32:10.063917336 +0000 UTC m=+152.651895602" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.132275 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lk7cs"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.140318 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.145621 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.159067 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.659043227 +0000 UTC m=+153.247021493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.185047 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.204355 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-59g9q" podStartSLOduration=133.204337705 podStartE2EDuration="2m13.204337705s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:10.200299581 +0000 UTC m=+152.788277847" watchObservedRunningTime="2025-11-28 14:32:10.204337705 +0000 UTC m=+152.792315971" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.204826 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lk7cs"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.257962 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-utilities\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.258050 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.258072 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-catalog-content\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.258091 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4gpl\" (UniqueName: \"kubernetes.io/projected/d0630627-d991-4c3b-ab3b-6a7711ea53f6-kube-api-access-q4gpl\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.258397 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.758385077 +0000 UTC m=+153.346363343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.290394 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ntw9s" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.325184 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wp6q4"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.326531 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" podStartSLOduration=133.326514212 podStartE2EDuration="2m13.326514212s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:10.324389757 +0000 UTC m=+152.912368013" watchObservedRunningTime="2025-11-28 14:32:10.326514212 +0000 UTC m=+152.914492478" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.327081 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.335867 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pknqn" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.344406 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.361667 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.361875 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-utilities\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.361918 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4gpl\" (UniqueName: \"kubernetes.io/projected/d0630627-d991-4c3b-ab3b-6a7711ea53f6-kube-api-access-q4gpl\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.361934 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-catalog-content\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.361961 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwdlp\" (UniqueName: \"kubernetes.io/projected/8789692c-b773-4938-8669-5f76d7017c14-kube-api-access-cwdlp\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.362000 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-utilities\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.362031 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-catalog-content\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.362143 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.86211931 +0000 UTC m=+153.450097566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.362684 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-catalog-content\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.362901 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-utilities\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.377010 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wp6q4"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.384696 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q42hh" podStartSLOduration=133.384673701 podStartE2EDuration="2m13.384673701s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:10.381231522 +0000 UTC m=+152.969209788" watchObservedRunningTime="2025-11-28 14:32:10.384673701 +0000 UTC m=+152.972651967" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.431768 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4gpl\" (UniqueName: \"kubernetes.io/projected/d0630627-d991-4c3b-ab3b-6a7711ea53f6-kube-api-access-q4gpl\") pod \"community-operators-lk7cs\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.451166 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" podStartSLOduration=133.451152004 podStartE2EDuration="2m13.451152004s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:10.44984468 +0000 UTC m=+153.037822946" watchObservedRunningTime="2025-11-28 14:32:10.451152004 +0000 UTC m=+153.039130270" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.463682 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.465413 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-utilities\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.465455 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.465488 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwdlp\" (UniqueName: \"kubernetes.io/projected/8789692c-b773-4938-8669-5f76d7017c14-kube-api-access-cwdlp\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.465540 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-catalog-content\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.465866 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:10.965854073 +0000 UTC m=+153.553832339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.465923 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-catalog-content\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.466011 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-utilities\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.492425 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" podStartSLOduration=133.492395367 podStartE2EDuration="2m13.492395367s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:10.492099289 +0000 UTC m=+153.080077555" watchObservedRunningTime="2025-11-28 14:32:10.492395367 +0000 UTC m=+153.080373633" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.507053 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwdlp\" (UniqueName: \"kubernetes.io/projected/8789692c-b773-4938-8669-5f76d7017c14-kube-api-access-cwdlp\") pod \"certified-operators-wp6q4\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.515965 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cwm9b"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.516821 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.546645 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cwm9b"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.568206 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.568407 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-catalog-content\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.568460 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-utilities\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.568545 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb666\" (UniqueName: \"kubernetes.io/projected/d691474c-0277-427e-abec-03f8df5fa5c1-kube-api-access-xb666\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.568669 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.068650882 +0000 UTC m=+153.656629148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.658011 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.670854 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb666\" (UniqueName: \"kubernetes.io/projected/d691474c-0277-427e-abec-03f8df5fa5c1-kube-api-access-xb666\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.670913 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-catalog-content\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.670951 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-utilities\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.671002 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.671256 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.171244575 +0000 UTC m=+153.759222841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.672027 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-catalog-content\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.672247 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-utilities\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.699459 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb666\" (UniqueName: \"kubernetes.io/projected/d691474c-0277-427e-abec-03f8df5fa5c1-kube-api-access-xb666\") pod \"community-operators-cwm9b\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.712590 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c27fw"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.713435 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.724359 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c27fw"] Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.772579 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.773133 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5ms\" (UniqueName: \"kubernetes.io/projected/5352e17f-a670-48d4-8ca0-52287875cd21-kube-api-access-lh5ms\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.773172 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-utilities\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.773187 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-catalog-content\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.773315 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.273297745 +0000 UTC m=+153.861276011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.844037 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.874616 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh5ms\" (UniqueName: \"kubernetes.io/projected/5352e17f-a670-48d4-8ca0-52287875cd21-kube-api-access-lh5ms\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.874659 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-utilities\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.874678 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-catalog-content\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.874699 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.875169 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.37515589 +0000 UTC m=+153.963134156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.875925 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-utilities\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.876144 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-catalog-content\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.890424 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:10 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:10 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:10 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.890497 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.896050 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh5ms\" (UniqueName: \"kubernetes.io/projected/5352e17f-a670-48d4-8ca0-52287875cd21-kube-api-access-lh5ms\") pod \"certified-operators-c27fw\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.925682 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lk7cs"] Nov 28 14:32:10 crc kubenswrapper[4817]: W1128 14:32:10.943566 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0630627_d991_4c3b_ab3b_6a7711ea53f6.slice/crio-80c570d47874d212a5df983d831c9959a2da9ea00eafdcf605391e85379a4eb8 WatchSource:0}: Error finding container 80c570d47874d212a5df983d831c9959a2da9ea00eafdcf605391e85379a4eb8: Status 404 returned error can't find the container with id 80c570d47874d212a5df983d831c9959a2da9ea00eafdcf605391e85379a4eb8 Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.979910 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.980029 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.479988931 +0000 UTC m=+154.067967207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.980470 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:10 crc kubenswrapper[4817]: E1128 14:32:10.980887 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.480870954 +0000 UTC m=+154.068849220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:10 crc kubenswrapper[4817]: I1128 14:32:10.984593 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wp6q4"] Nov 28 14:32:10 crc kubenswrapper[4817]: W1128 14:32:10.999497 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8789692c_b773_4938_8669_5f76d7017c14.slice/crio-c2039329da8d204f428430342348e1867f2ff8fbd8be6feb61e1681878d3280a WatchSource:0}: Error finding container c2039329da8d204f428430342348e1867f2ff8fbd8be6feb61e1681878d3280a: Status 404 returned error can't find the container with id c2039329da8d204f428430342348e1867f2ff8fbd8be6feb61e1681878d3280a Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.039206 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp6q4" event={"ID":"8789692c-b773-4938-8669-5f76d7017c14","Type":"ContainerStarted","Data":"c2039329da8d204f428430342348e1867f2ff8fbd8be6feb61e1681878d3280a"} Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.040649 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk7cs" event={"ID":"d0630627-d991-4c3b-ab3b-6a7711ea53f6","Type":"ContainerStarted","Data":"80c570d47874d212a5df983d831c9959a2da9ea00eafdcf605391e85379a4eb8"} Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.044820 4817 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-s4gdr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.044863 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" podUID="8b50b274-12b2-4ed3-ad0a-81365aeff72c" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.056302 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.068890 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cwm9b"] Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.081164 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.081344 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.581317542 +0000 UTC m=+154.169295818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.081965 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.084595 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.584582347 +0000 UTC m=+154.172560613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: W1128 14:32:11.085118 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd691474c_0277_427e_abec_03f8df5fa5c1.slice/crio-e391416a92847cc7aa16e6a726c3082b16e3de1f52ed40a85d59198a8bef9b76 WatchSource:0}: Error finding container e391416a92847cc7aa16e6a726c3082b16e3de1f52ed40a85d59198a8bef9b76: Status 404 returned error can't find the container with id e391416a92847cc7aa16e6a726c3082b16e3de1f52ed40a85d59198a8bef9b76 Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.183424 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.183532 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.683513806 +0000 UTC m=+154.271492072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.183946 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.683938177 +0000 UTC m=+154.271916443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.184082 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.284826 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.285207 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.785191796 +0000 UTC m=+154.373170062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.288339 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.315481 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c27fw"] Nov 28 14:32:11 crc kubenswrapper[4817]: W1128 14:32:11.324817 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5352e17f_a670_48d4_8ca0_52287875cd21.slice/crio-55e1f5c9341f6cbad3141870a7ec9fe2a20bb1b505e90d9a019e437c68b18469 WatchSource:0}: Error finding container 55e1f5c9341f6cbad3141870a7ec9fe2a20bb1b505e90d9a019e437c68b18469: Status 404 returned error can't find the container with id 55e1f5c9341f6cbad3141870a7ec9fe2a20bb1b505e90d9a019e437c68b18469 Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.386276 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wm9p\" (UniqueName: \"kubernetes.io/projected/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-kube-api-access-6wm9p\") pod \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.386457 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-secret-volume\") pod \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.386633 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-config-volume\") pod \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\" (UID: \"d52d05b7-f5b8-4fb7-bd09-e063e48c615d\") " Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.386816 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.387080 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.887067441 +0000 UTC m=+154.475045707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.387663 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-config-volume" (OuterVolumeSpecName: "config-volume") pod "d52d05b7-f5b8-4fb7-bd09-e063e48c615d" (UID: "d52d05b7-f5b8-4fb7-bd09-e063e48c615d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.392406 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-kube-api-access-6wm9p" (OuterVolumeSpecName: "kube-api-access-6wm9p") pod "d52d05b7-f5b8-4fb7-bd09-e063e48c615d" (UID: "d52d05b7-f5b8-4fb7-bd09-e063e48c615d"). InnerVolumeSpecName "kube-api-access-6wm9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.395856 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d52d05b7-f5b8-4fb7-bd09-e063e48c615d" (UID: "d52d05b7-f5b8-4fb7-bd09-e063e48c615d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.487531 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.487911 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.487925 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wm9p\" (UniqueName: \"kubernetes.io/projected/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-kube-api-access-6wm9p\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.487935 4817 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d52d05b7-f5b8-4fb7-bd09-e063e48c615d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.487998 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:11.987981092 +0000 UTC m=+154.575959358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.588706 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.588983 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:12.088968194 +0000 UTC m=+154.676946460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.689920 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.690254 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:12.190239724 +0000 UTC m=+154.778217990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.725136 4817 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.791640 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.792018 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:12.292002376 +0000 UTC m=+154.879980642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.887487 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:11 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:11 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:11 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.887762 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.892650 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.892836 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 14:32:12.392814674 +0000 UTC m=+154.980792940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.893006 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:11 crc kubenswrapper[4817]: E1128 14:32:11.893393 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 14:32:12.393377199 +0000 UTC m=+154.981355465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v76hp" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.898514 4817 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-28T14:32:11.725166204Z","Handler":null,"Name":""} Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.900925 4817 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.900957 4817 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 28 14:32:11 crc kubenswrapper[4817]: I1128 14:32:11.994019 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.004365 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.046292 4817 generic.go:334] "Generic (PLEG): container finished" podID="8789692c-b773-4938-8669-5f76d7017c14" containerID="8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76" exitCode=0 Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.046388 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp6q4" event={"ID":"8789692c-b773-4938-8669-5f76d7017c14","Type":"ContainerDied","Data":"8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.047746 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.048076 4817 generic.go:334] "Generic (PLEG): container finished" podID="5352e17f-a670-48d4-8ca0-52287875cd21" containerID="e49a9281d168440d3cde0b6c358fa79a433afec381e9076f8f091c36f89952ef" exitCode=0 Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.048148 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c27fw" event={"ID":"5352e17f-a670-48d4-8ca0-52287875cd21","Type":"ContainerDied","Data":"e49a9281d168440d3cde0b6c358fa79a433afec381e9076f8f091c36f89952ef"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.048173 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c27fw" event={"ID":"5352e17f-a670-48d4-8ca0-52287875cd21","Type":"ContainerStarted","Data":"55e1f5c9341f6cbad3141870a7ec9fe2a20bb1b505e90d9a019e437c68b18469"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.074895 4817 generic.go:334] "Generic (PLEG): container finished" podID="d691474c-0277-427e-abec-03f8df5fa5c1" containerID="30eb8390d1e8ff0d4b3ff57988f63db08dec69b6b5389419e71a4a7dacc1eb1e" exitCode=0 Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.075302 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwm9b" event={"ID":"d691474c-0277-427e-abec-03f8df5fa5c1","Type":"ContainerDied","Data":"30eb8390d1e8ff0d4b3ff57988f63db08dec69b6b5389419e71a4a7dacc1eb1e"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.075348 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwm9b" event={"ID":"d691474c-0277-427e-abec-03f8df5fa5c1","Type":"ContainerStarted","Data":"e391416a92847cc7aa16e6a726c3082b16e3de1f52ed40a85d59198a8bef9b76"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.091568 4817 generic.go:334] "Generic (PLEG): container finished" podID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerID="989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3" exitCode=0 Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.091627 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk7cs" event={"ID":"d0630627-d991-4c3b-ab3b-6a7711ea53f6","Type":"ContainerDied","Data":"989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.097050 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.102984 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mj798" event={"ID":"7bc1102b-0f53-4b51-b5c4-eeb118249dea","Type":"ContainerStarted","Data":"a718d217952fdd8f233bfe08f281c7b5f371b0d5e470deac8c8b19fabfbdd155"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.103039 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mj798" event={"ID":"7bc1102b-0f53-4b51-b5c4-eeb118249dea","Type":"ContainerStarted","Data":"a97a4c191a5d9d5e9db5851122b9e7ca91c1a5e9d14e38a5621d941b7e4fdee3"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.107401 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.107533 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx" event={"ID":"d52d05b7-f5b8-4fb7-bd09-e063e48c615d","Type":"ContainerDied","Data":"4192e7b6cc583c2a1feeb60a76fe46232ad86eac6478fceeb3252aff789f1403"} Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.107570 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4192e7b6cc583c2a1feeb60a76fe46232ad86eac6478fceeb3252aff789f1403" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.113310 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.113375 4817 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.113408 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.147842 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v76hp\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.179241 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.248389 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.303156 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c4l97"] Nov 28 14:32:12 crc kubenswrapper[4817]: E1128 14:32:12.303358 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d52d05b7-f5b8-4fb7-bd09-e063e48c615d" containerName="collect-profiles" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.303369 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d52d05b7-f5b8-4fb7-bd09-e063e48c615d" containerName="collect-profiles" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.303504 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d52d05b7-f5b8-4fb7-bd09-e063e48c615d" containerName="collect-profiles" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.304165 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.309090 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.314331 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4l97"] Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.408599 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sdcb\" (UniqueName: \"kubernetes.io/projected/041e51f8-a1ce-41a6-adba-b0b8bd638de3-kube-api-access-9sdcb\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.408638 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-catalog-content\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.408681 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-utilities\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.509377 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-catalog-content\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.509991 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-utilities\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.510070 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sdcb\" (UniqueName: \"kubernetes.io/projected/041e51f8-a1ce-41a6-adba-b0b8bd638de3-kube-api-access-9sdcb\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.512437 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-catalog-content\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.513486 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-utilities\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.531474 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sdcb\" (UniqueName: \"kubernetes.io/projected/041e51f8-a1ce-41a6-adba-b0b8bd638de3-kube-api-access-9sdcb\") pod \"redhat-marketplace-c4l97\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.565478 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v76hp"] Nov 28 14:32:12 crc kubenswrapper[4817]: W1128 14:32:12.573538 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0ef5080_e168_4e13_92d5_8a41f51610da.slice/crio-d89754a5ab79e371c526aeb9967fe8683b3b5ed7dc7b5349241d1722303f4c0d WatchSource:0}: Error finding container d89754a5ab79e371c526aeb9967fe8683b3b5ed7dc7b5349241d1722303f4c0d: Status 404 returned error can't find the container with id d89754a5ab79e371c526aeb9967fe8683b3b5ed7dc7b5349241d1722303f4c0d Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.628020 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.703309 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.704945 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.721549 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.815788 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-utilities\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.815903 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svf2f\" (UniqueName: \"kubernetes.io/projected/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-kube-api-access-svf2f\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.815949 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-catalog-content\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.837180 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4l97"] Nov 28 14:32:12 crc kubenswrapper[4817]: W1128 14:32:12.850711 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod041e51f8_a1ce_41a6_adba_b0b8bd638de3.slice/crio-3873eb7585c9fe8d98c6ddaa92b9e68292e644e6ca134c02300aab45c6796f0f WatchSource:0}: Error finding container 3873eb7585c9fe8d98c6ddaa92b9e68292e644e6ca134c02300aab45c6796f0f: Status 404 returned error can't find the container with id 3873eb7585c9fe8d98c6ddaa92b9e68292e644e6ca134c02300aab45c6796f0f Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.887647 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:12 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:12 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:12 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.887710 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.918756 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svf2f\" (UniqueName: \"kubernetes.io/projected/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-kube-api-access-svf2f\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.918873 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-catalog-content\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.918931 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-utilities\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.919386 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-catalog-content\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.920670 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-utilities\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:12 crc kubenswrapper[4817]: I1128 14:32:12.938495 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svf2f\" (UniqueName: \"kubernetes.io/projected/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-kube-api-access-svf2f\") pod \"redhat-marketplace-phfbn\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.028001 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.113673 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mj798" event={"ID":"7bc1102b-0f53-4b51-b5c4-eeb118249dea","Type":"ContainerStarted","Data":"08f3da240c2a9cf2e5340046a045b1c891df6d70bc3243ad37d48e8d275a01c1"} Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.119111 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" event={"ID":"f0ef5080-e168-4e13-92d5-8a41f51610da","Type":"ContainerStarted","Data":"2b88bf25abf875fba580d01c5232e33f33f772f716e39877873af13b367f89a7"} Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.119161 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" event={"ID":"f0ef5080-e168-4e13-92d5-8a41f51610da","Type":"ContainerStarted","Data":"d89754a5ab79e371c526aeb9967fe8683b3b5ed7dc7b5349241d1722303f4c0d"} Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.119822 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.126035 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4l97" event={"ID":"041e51f8-a1ce-41a6-adba-b0b8bd638de3","Type":"ContainerStarted","Data":"3873eb7585c9fe8d98c6ddaa92b9e68292e644e6ca134c02300aab45c6796f0f"} Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.140393 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-mj798" podStartSLOduration=12.140373072 podStartE2EDuration="12.140373072s" podCreationTimestamp="2025-11-28 14:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:13.138691529 +0000 UTC m=+155.726669805" watchObservedRunningTime="2025-11-28 14:32:13.140373072 +0000 UTC m=+155.728351338" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.166796 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" podStartSLOduration=136.166774433 podStartE2EDuration="2m16.166774433s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:13.160556752 +0000 UTC m=+155.748535018" watchObservedRunningTime="2025-11-28 14:32:13.166774433 +0000 UTC m=+155.754752699" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.302373 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-69p7n"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.304986 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.316397 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.325270 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-catalog-content\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.325330 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkqdt\" (UniqueName: \"kubernetes.io/projected/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-kube-api-access-mkqdt\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.325373 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-utilities\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.330772 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69p7n"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.426433 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkqdt\" (UniqueName: \"kubernetes.io/projected/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-kube-api-access-mkqdt\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.426522 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-utilities\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.426581 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-catalog-content\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.427234 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-catalog-content\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.427373 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-utilities\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.467094 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkqdt\" (UniqueName: \"kubernetes.io/projected/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-kube-api-access-mkqdt\") pod \"redhat-operators-69p7n\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.475736 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.633570 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.698790 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tsnzb"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.699920 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.706252 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tsnzb"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.730734 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-catalog-content\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.730805 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-utilities\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.730828 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqc6r\" (UniqueName: \"kubernetes.io/projected/690548fe-586d-479d-b937-f47a312ef236-kube-api-access-tqc6r\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.745025 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.793881 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.794525 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.796143 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.798387 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.806834 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.831832 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-catalog-content\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.831918 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.831950 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.832082 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-utilities\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.832114 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqc6r\" (UniqueName: \"kubernetes.io/projected/690548fe-586d-479d-b937-f47a312ef236-kube-api-access-tqc6r\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.833317 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-catalog-content\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.833354 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-utilities\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.848301 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqc6r\" (UniqueName: \"kubernetes.io/projected/690548fe-586d-479d-b937-f47a312ef236-kube-api-access-tqc6r\") pod \"redhat-operators-tsnzb\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.889230 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:13 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:13 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:13 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.889406 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.905874 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69p7n"] Nov 28 14:32:13 crc kubenswrapper[4817]: W1128 14:32:13.917539 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76973893_1cf3_40bd_aa5e_e2a4a32ec10a.slice/crio-ff05589226b87cbe3b31a49d34f9196a00ab2c3e440bf3aaa93437108a309050 WatchSource:0}: Error finding container ff05589226b87cbe3b31a49d34f9196a00ab2c3e440bf3aaa93437108a309050: Status 404 returned error can't find the container with id ff05589226b87cbe3b31a49d34f9196a00ab2c3e440bf3aaa93437108a309050 Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.937181 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.937432 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.937484 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:13 crc kubenswrapper[4817]: I1128 14:32:13.954078 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.006236 4817 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kzwm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.006288 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5kzwm" podUID="8e94a88b-c6a1-4edd-8759-136275441876" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.006433 4817 patch_prober.go:28] interesting pod/downloads-7954f5f757-5kzwm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.006513 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5kzwm" podUID="8e94a88b-c6a1-4edd-8759-136275441876" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.019570 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.055927 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.056156 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.057045 4817 patch_prober.go:28] interesting pod/console-f9d7485db-9xjrk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.057089 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9xjrk" podUID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.124297 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.133984 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerStarted","Data":"4b5f6e28647041c823e788c6b6399d6381c81fb66c3483cc66fbb70a6780fbe4"} Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.141486 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p7n" event={"ID":"76973893-1cf3-40bd-aa5e-e2a4a32ec10a","Type":"ContainerStarted","Data":"ff05589226b87cbe3b31a49d34f9196a00ab2c3e440bf3aaa93437108a309050"} Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.146940 4817 generic.go:334] "Generic (PLEG): container finished" podID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerID="79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731" exitCode=0 Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.147036 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4l97" event={"ID":"041e51f8-a1ce-41a6-adba-b0b8bd638de3","Type":"ContainerDied","Data":"79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731"} Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.162966 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.163528 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.176404 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.221884 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.222341 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.243463 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.448649 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 14:32:14 crc kubenswrapper[4817]: W1128 14:32:14.454606 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod51d0d90f_647c_4bc2_bc06_1d3a89aafafb.slice/crio-3b45b6c9694a8bd7a53e94e579748e4cd6053d2cbc9252c96a92125e95f85162 WatchSource:0}: Error finding container 3b45b6c9694a8bd7a53e94e579748e4cd6053d2cbc9252c96a92125e95f85162: Status 404 returned error can't find the container with id 3b45b6c9694a8bd7a53e94e579748e4cd6053d2cbc9252c96a92125e95f85162 Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.541892 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tsnzb"] Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.886004 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.892334 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:14 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:14 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:14 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:14 crc kubenswrapper[4817]: I1128 14:32:14.892404 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.185983 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51d0d90f-647c-4bc2-bc06-1d3a89aafafb","Type":"ContainerStarted","Data":"7dbd441e71957f68754f34e1758d3c81daebf8d450a0febb18b9301ddc93c5fc"} Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.186051 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51d0d90f-647c-4bc2-bc06-1d3a89aafafb","Type":"ContainerStarted","Data":"3b45b6c9694a8bd7a53e94e579748e4cd6053d2cbc9252c96a92125e95f85162"} Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.188282 4817 generic.go:334] "Generic (PLEG): container finished" podID="690548fe-586d-479d-b937-f47a312ef236" containerID="2c4b9c980bfc8d9c306e82d4d7ac400ca99629968e2b73ead39bba377d33d5a0" exitCode=0 Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.188633 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsnzb" event={"ID":"690548fe-586d-479d-b937-f47a312ef236","Type":"ContainerDied","Data":"2c4b9c980bfc8d9c306e82d4d7ac400ca99629968e2b73ead39bba377d33d5a0"} Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.188712 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsnzb" event={"ID":"690548fe-586d-479d-b937-f47a312ef236","Type":"ContainerStarted","Data":"fcaa4a7d476cf8e869ff802aaf3efc1ab4cc9402099d61a7ebebad9fcf2ccded"} Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.196409 4817 generic.go:334] "Generic (PLEG): container finished" podID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerID="e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc" exitCode=0 Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.196524 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p7n" event={"ID":"76973893-1cf3-40bd-aa5e-e2a4a32ec10a","Type":"ContainerDied","Data":"e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc"} Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.201178 4817 generic.go:334] "Generic (PLEG): container finished" podID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerID="9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9" exitCode=0 Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.201323 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerDied","Data":"9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9"} Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.206306 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-v48xp" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.212139 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-27xnz" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.215689 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.215653409 podStartE2EDuration="2.215653409s" podCreationTimestamp="2025-11-28 14:32:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:15.198415175 +0000 UTC m=+157.786393441" watchObservedRunningTime="2025-11-28 14:32:15.215653409 +0000 UTC m=+157.803631665" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.498869 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.503127 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.510667 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.511357 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.512455 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.570612 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.570706 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.671947 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.672033 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.672117 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.707385 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.829977 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.887769 4817 patch_prober.go:28] interesting pod/router-default-5444994796-82mjn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 14:32:15 crc kubenswrapper[4817]: [-]has-synced failed: reason withheld Nov 28 14:32:15 crc kubenswrapper[4817]: [+]process-running ok Nov 28 14:32:15 crc kubenswrapper[4817]: healthz check failed Nov 28 14:32:15 crc kubenswrapper[4817]: I1128 14:32:15.887893 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-82mjn" podUID="b29cc89d-c829-41b8-bbb2-5a35b55ad6b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 14:32:16 crc kubenswrapper[4817]: I1128 14:32:16.120081 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 14:32:16 crc kubenswrapper[4817]: W1128 14:32:16.140024 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5fcfddb9_9e2f_4d86_96f2_a21c13ee49c8.slice/crio-b76e6f05a6220e3c96f18c30244a85da8cb3343ca370cd3f4c11ccdddb066d67 WatchSource:0}: Error finding container b76e6f05a6220e3c96f18c30244a85da8cb3343ca370cd3f4c11ccdddb066d67: Status 404 returned error can't find the container with id b76e6f05a6220e3c96f18c30244a85da8cb3343ca370cd3f4c11ccdddb066d67 Nov 28 14:32:16 crc kubenswrapper[4817]: I1128 14:32:16.208617 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8","Type":"ContainerStarted","Data":"b76e6f05a6220e3c96f18c30244a85da8cb3343ca370cd3f4c11ccdddb066d67"} Nov 28 14:32:16 crc kubenswrapper[4817]: I1128 14:32:16.212001 4817 generic.go:334] "Generic (PLEG): container finished" podID="51d0d90f-647c-4bc2-bc06-1d3a89aafafb" containerID="7dbd441e71957f68754f34e1758d3c81daebf8d450a0febb18b9301ddc93c5fc" exitCode=0 Nov 28 14:32:16 crc kubenswrapper[4817]: I1128 14:32:16.212091 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51d0d90f-647c-4bc2-bc06-1d3a89aafafb","Type":"ContainerDied","Data":"7dbd441e71957f68754f34e1758d3c81daebf8d450a0febb18b9301ddc93c5fc"} Nov 28 14:32:16 crc kubenswrapper[4817]: I1128 14:32:16.900476 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:16 crc kubenswrapper[4817]: I1128 14:32:16.904706 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-82mjn" Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.484635 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.614512 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kubelet-dir\") pod \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.614886 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kube-api-access\") pod \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\" (UID: \"51d0d90f-647c-4bc2-bc06-1d3a89aafafb\") " Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.614656 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "51d0d90f-647c-4bc2-bc06-1d3a89aafafb" (UID: "51d0d90f-647c-4bc2-bc06-1d3a89aafafb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.615155 4817 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.621287 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "51d0d90f-647c-4bc2-bc06-1d3a89aafafb" (UID: "51d0d90f-647c-4bc2-bc06-1d3a89aafafb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:32:17 crc kubenswrapper[4817]: I1128 14:32:17.729167 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d0d90f-647c-4bc2-bc06-1d3a89aafafb-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:18 crc kubenswrapper[4817]: I1128 14:32:18.226376 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8","Type":"ContainerStarted","Data":"dc40e69f6b7ca9a5fda89549354e5d2840814e62fdab0653ae6b2480294f45e0"} Nov 28 14:32:18 crc kubenswrapper[4817]: I1128 14:32:18.230728 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51d0d90f-647c-4bc2-bc06-1d3a89aafafb","Type":"ContainerDied","Data":"3b45b6c9694a8bd7a53e94e579748e4cd6053d2cbc9252c96a92125e95f85162"} Nov 28 14:32:18 crc kubenswrapper[4817]: I1128 14:32:18.230776 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b45b6c9694a8bd7a53e94e579748e4cd6053d2cbc9252c96a92125e95f85162" Nov 28 14:32:18 crc kubenswrapper[4817]: I1128 14:32:18.230803 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 14:32:18 crc kubenswrapper[4817]: I1128 14:32:18.240415 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.240396153 podStartE2EDuration="3.240396153s" podCreationTimestamp="2025-11-28 14:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:18.238513064 +0000 UTC m=+160.826491330" watchObservedRunningTime="2025-11-28 14:32:18.240396153 +0000 UTC m=+160.828374419" Nov 28 14:32:19 crc kubenswrapper[4817]: I1128 14:32:19.241846 4817 generic.go:334] "Generic (PLEG): container finished" podID="5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8" containerID="dc40e69f6b7ca9a5fda89549354e5d2840814e62fdab0653ae6b2480294f45e0" exitCode=0 Nov 28 14:32:19 crc kubenswrapper[4817]: I1128 14:32:19.241890 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8","Type":"ContainerDied","Data":"dc40e69f6b7ca9a5fda89549354e5d2840814e62fdab0653ae6b2480294f45e0"} Nov 28 14:32:19 crc kubenswrapper[4817]: I1128 14:32:19.588914 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:32:19 crc kubenswrapper[4817]: I1128 14:32:19.599126 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bfd936ce-70b3-4e00-944f-f505238d1ab2-metrics-certs\") pod \"network-metrics-daemon-zvnxb\" (UID: \"bfd936ce-70b3-4e00-944f-f505238d1ab2\") " pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:32:19 crc kubenswrapper[4817]: I1128 14:32:19.854781 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zvnxb" Nov 28 14:32:19 crc kubenswrapper[4817]: I1128 14:32:19.918370 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-mbk86" Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.132823 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zvnxb"] Nov 28 14:32:20 crc kubenswrapper[4817]: W1128 14:32:20.143210 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfd936ce_70b3_4e00_944f_f505238d1ab2.slice/crio-cd0a9e376e8a696d876b61a28dff8b2744cf17c73fcd34da12b9bc0619dac544 WatchSource:0}: Error finding container cd0a9e376e8a696d876b61a28dff8b2744cf17c73fcd34da12b9bc0619dac544: Status 404 returned error can't find the container with id cd0a9e376e8a696d876b61a28dff8b2744cf17c73fcd34da12b9bc0619dac544 Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.250243 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" event={"ID":"bfd936ce-70b3-4e00-944f-f505238d1ab2","Type":"ContainerStarted","Data":"cd0a9e376e8a696d876b61a28dff8b2744cf17c73fcd34da12b9bc0619dac544"} Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.499865 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.603267 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kubelet-dir\") pod \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.603576 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kube-api-access\") pod \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\" (UID: \"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8\") " Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.603380 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8" (UID: "5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.603866 4817 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.608859 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8" (UID: "5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:32:20 crc kubenswrapper[4817]: I1128 14:32:20.705436 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:32:21 crc kubenswrapper[4817]: I1128 14:32:21.257280 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8","Type":"ContainerDied","Data":"b76e6f05a6220e3c96f18c30244a85da8cb3343ca370cd3f4c11ccdddb066d67"} Nov 28 14:32:21 crc kubenswrapper[4817]: I1128 14:32:21.257325 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 14:32:21 crc kubenswrapper[4817]: I1128 14:32:21.257332 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b76e6f05a6220e3c96f18c30244a85da8cb3343ca370cd3f4c11ccdddb066d67" Nov 28 14:32:22 crc kubenswrapper[4817]: I1128 14:32:22.262558 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" event={"ID":"bfd936ce-70b3-4e00-944f-f505238d1ab2","Type":"ContainerStarted","Data":"5db9f43f655a13f311938706609875a5ac50424a101baf60b72f9b42bdf8e40d"} Nov 28 14:32:24 crc kubenswrapper[4817]: I1128 14:32:24.018879 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5kzwm" Nov 28 14:32:24 crc kubenswrapper[4817]: I1128 14:32:24.047623 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:32:24 crc kubenswrapper[4817]: I1128 14:32:24.047698 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:32:24 crc kubenswrapper[4817]: I1128 14:32:24.059546 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:24 crc kubenswrapper[4817]: I1128 14:32:24.063495 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:32:32 crc kubenswrapper[4817]: I1128 14:32:32.253131 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:32:44 crc kubenswrapper[4817]: I1128 14:32:44.521884 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cx5k6" Nov 28 14:32:44 crc kubenswrapper[4817]: I1128 14:32:44.736400 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 14:32:48 crc kubenswrapper[4817]: E1128 14:32:48.042279 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 14:32:48 crc kubenswrapper[4817]: E1128 14:32:48.042763 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cwdlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wp6q4_openshift-marketplace(8789692c-b773-4938-8669-5f76d7017c14): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:48 crc kubenswrapper[4817]: E1128 14:32:48.044494 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wp6q4" podUID="8789692c-b773-4938-8669-5f76d7017c14" Nov 28 14:32:49 crc kubenswrapper[4817]: E1128 14:32:49.948481 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wp6q4" podUID="8789692c-b773-4938-8669-5f76d7017c14" Nov 28 14:32:51 crc kubenswrapper[4817]: E1128 14:32:51.310842 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 14:32:51 crc kubenswrapper[4817]: E1128 14:32:51.311097 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q4gpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lk7cs_openshift-marketplace(d0630627-d991-4c3b-ab3b-6a7711ea53f6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:51 crc kubenswrapper[4817]: E1128 14:32:51.312564 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lk7cs" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" Nov 28 14:32:51 crc kubenswrapper[4817]: E1128 14:32:51.640909 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lk7cs" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" Nov 28 14:32:54 crc kubenswrapper[4817]: I1128 14:32:54.045162 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:32:54 crc kubenswrapper[4817]: I1128 14:32:54.045248 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:32:54 crc kubenswrapper[4817]: E1128 14:32:54.601657 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 14:32:54 crc kubenswrapper[4817]: E1128 14:32:54.602023 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9sdcb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-c4l97_openshift-marketplace(041e51f8-a1ce-41a6-adba-b0b8bd638de3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:54 crc kubenswrapper[4817]: E1128 14:32:54.603253 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-c4l97" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" Nov 28 14:32:54 crc kubenswrapper[4817]: E1128 14:32:54.695806 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 14:32:54 crc kubenswrapper[4817]: E1128 14:32:54.696021 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lh5ms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-c27fw_openshift-marketplace(5352e17f-a670-48d4-8ca0-52287875cd21): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:54 crc kubenswrapper[4817]: E1128 14:32:54.697376 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-c27fw" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.584305 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 14:32:55 crc kubenswrapper[4817]: E1128 14:32:55.584579 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51d0d90f-647c-4bc2-bc06-1d3a89aafafb" containerName="pruner" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.584596 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="51d0d90f-647c-4bc2-bc06-1d3a89aafafb" containerName="pruner" Nov 28 14:32:55 crc kubenswrapper[4817]: E1128 14:32:55.584609 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8" containerName="pruner" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.584618 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8" containerName="pruner" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.584765 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="51d0d90f-647c-4bc2-bc06-1d3a89aafafb" containerName="pruner" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.584787 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fcfddb9-9e2f-4d86-96f2-a21c13ee49c8" containerName="pruner" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.585269 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.605477 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.606957 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.617315 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.707063 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.707121 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.808116 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.808198 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.808278 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.826755 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:55 crc kubenswrapper[4817]: I1128 14:32:55.921405 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.685253 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-c27fw" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.685271 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-c4l97" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.709579 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.709757 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-svf2f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-phfbn_openshift-marketplace(5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.710637 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.710773 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqc6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tsnzb_openshift-marketplace(690548fe-586d-479d-b937-f47a312ef236): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.711813 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-phfbn" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.711917 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tsnzb" podUID="690548fe-586d-479d-b937-f47a312ef236" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.714193 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.714431 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mkqdt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-69p7n_openshift-marketplace(76973893-1cf3-40bd-aa5e-e2a4a32ec10a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.715667 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-69p7n" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.793875 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.794285 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xb666,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cwm9b_openshift-marketplace(d691474c-0277-427e-abec-03f8df5fa5c1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 14:32:57 crc kubenswrapper[4817]: E1128 14:32:57.795578 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cwm9b" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" Nov 28 14:32:58 crc kubenswrapper[4817]: I1128 14:32:58.110629 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 14:32:58 crc kubenswrapper[4817]: W1128 14:32:58.121361 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3c8e3fd7_d484_4cf1_8b47_fc11dd51e8a7.slice/crio-47963f632e01e545b94f654954fbe1352a7eba678d2b0dd814d22d09993386b0 WatchSource:0}: Error finding container 47963f632e01e545b94f654954fbe1352a7eba678d2b0dd814d22d09993386b0: Status 404 returned error can't find the container with id 47963f632e01e545b94f654954fbe1352a7eba678d2b0dd814d22d09993386b0 Nov 28 14:32:58 crc kubenswrapper[4817]: I1128 14:32:58.485250 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zvnxb" event={"ID":"bfd936ce-70b3-4e00-944f-f505238d1ab2","Type":"ContainerStarted","Data":"da3bc345a54503bc77195334d1af04fc834823a88720b0e81ed7cc699a05be8a"} Nov 28 14:32:58 crc kubenswrapper[4817]: I1128 14:32:58.486801 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7","Type":"ContainerStarted","Data":"47963f632e01e545b94f654954fbe1352a7eba678d2b0dd814d22d09993386b0"} Nov 28 14:32:58 crc kubenswrapper[4817]: E1128 14:32:58.488703 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cwm9b" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" Nov 28 14:32:58 crc kubenswrapper[4817]: E1128 14:32:58.489018 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-69p7n" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" Nov 28 14:32:58 crc kubenswrapper[4817]: E1128 14:32:58.489090 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-phfbn" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" Nov 28 14:32:58 crc kubenswrapper[4817]: E1128 14:32:58.491037 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tsnzb" podUID="690548fe-586d-479d-b937-f47a312ef236" Nov 28 14:32:58 crc kubenswrapper[4817]: I1128 14:32:58.502814 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zvnxb" podStartSLOduration=181.502784215 podStartE2EDuration="3m1.502784215s" podCreationTimestamp="2025-11-28 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:32:58.500919346 +0000 UTC m=+201.088897622" watchObservedRunningTime="2025-11-28 14:32:58.502784215 +0000 UTC m=+201.090762521" Nov 28 14:32:59 crc kubenswrapper[4817]: I1128 14:32:59.495862 4817 generic.go:334] "Generic (PLEG): container finished" podID="3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7" containerID="6df726b2b25b7435937ccd77757f6e394345ac3a72a39836f7336098b5969b4b" exitCode=0 Nov 28 14:32:59 crc kubenswrapper[4817]: I1128 14:32:59.495962 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7","Type":"ContainerDied","Data":"6df726b2b25b7435937ccd77757f6e394345ac3a72a39836f7336098b5969b4b"} Nov 28 14:33:00 crc kubenswrapper[4817]: I1128 14:33:00.843499 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.017066 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kube-api-access\") pod \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.017180 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kubelet-dir\") pod \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\" (UID: \"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7\") " Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.017431 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7" (UID: "3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.025437 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7" (UID: "3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.119306 4817 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.119363 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.510312 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7","Type":"ContainerDied","Data":"47963f632e01e545b94f654954fbe1352a7eba678d2b0dd814d22d09993386b0"} Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.510366 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47963f632e01e545b94f654954fbe1352a7eba678d2b0dd814d22d09993386b0" Nov 28 14:33:01 crc kubenswrapper[4817]: I1128 14:33:01.510433 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.519091 4817 generic.go:334] "Generic (PLEG): container finished" podID="8789692c-b773-4938-8669-5f76d7017c14" containerID="4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a" exitCode=0 Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.519160 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp6q4" event={"ID":"8789692c-b773-4938-8669-5f76d7017c14","Type":"ContainerDied","Data":"4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a"} Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.778242 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 14:33:02 crc kubenswrapper[4817]: E1128 14:33:02.778830 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7" containerName="pruner" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.778849 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7" containerName="pruner" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.779055 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8e3fd7-d484-4cf1-8b47-fc11dd51e8a7" containerName="pruner" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.779571 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.782173 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.782241 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.786687 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.855599 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.855634 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-var-lock\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.855750 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kube-api-access\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.957179 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kube-api-access\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.957275 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.957302 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-var-lock\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.957434 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-var-lock\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.957834 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:02 crc kubenswrapper[4817]: I1128 14:33:02.982850 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kube-api-access\") pod \"installer-9-crc\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:03 crc kubenswrapper[4817]: I1128 14:33:03.099038 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:03 crc kubenswrapper[4817]: I1128 14:33:03.519576 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 14:33:03 crc kubenswrapper[4817]: W1128 14:33:03.536152 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod50ce10d4_a7fa_491f_bbe2_50117f74f86c.slice/crio-68e89bcde1d2fb28421124843fb29c26f98328a1c36ab8d3c7229f88ca2fb985 WatchSource:0}: Error finding container 68e89bcde1d2fb28421124843fb29c26f98328a1c36ab8d3c7229f88ca2fb985: Status 404 returned error can't find the container with id 68e89bcde1d2fb28421124843fb29c26f98328a1c36ab8d3c7229f88ca2fb985 Nov 28 14:33:04 crc kubenswrapper[4817]: I1128 14:33:04.531223 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"50ce10d4-a7fa-491f-bbe2-50117f74f86c","Type":"ContainerStarted","Data":"c8a83dbb5b959e70abbae07120c957a37a7bd5c4d96cc2ef314a5ba1c9fce26b"} Nov 28 14:33:04 crc kubenswrapper[4817]: I1128 14:33:04.531764 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"50ce10d4-a7fa-491f-bbe2-50117f74f86c","Type":"ContainerStarted","Data":"68e89bcde1d2fb28421124843fb29c26f98328a1c36ab8d3c7229f88ca2fb985"} Nov 28 14:33:04 crc kubenswrapper[4817]: I1128 14:33:04.535787 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp6q4" event={"ID":"8789692c-b773-4938-8669-5f76d7017c14","Type":"ContainerStarted","Data":"c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd"} Nov 28 14:33:04 crc kubenswrapper[4817]: I1128 14:33:04.567841 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wp6q4" podStartSLOduration=3.504578841 podStartE2EDuration="54.567822685s" podCreationTimestamp="2025-11-28 14:32:10 +0000 UTC" firstStartedPulling="2025-11-28 14:32:12.04749792 +0000 UTC m=+154.635476186" lastFinishedPulling="2025-11-28 14:33:03.110741764 +0000 UTC m=+205.698720030" observedRunningTime="2025-11-28 14:33:04.566630864 +0000 UTC m=+207.154609140" watchObservedRunningTime="2025-11-28 14:33:04.567822685 +0000 UTC m=+207.155800951" Nov 28 14:33:04 crc kubenswrapper[4817]: I1128 14:33:04.568083 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.568077321 podStartE2EDuration="2.568077321s" podCreationTimestamp="2025-11-28 14:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:33:04.54991229 +0000 UTC m=+207.137890556" watchObservedRunningTime="2025-11-28 14:33:04.568077321 +0000 UTC m=+207.156055597" Nov 28 14:33:10 crc kubenswrapper[4817]: I1128 14:33:10.569520 4817 generic.go:334] "Generic (PLEG): container finished" podID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerID="b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba" exitCode=0 Nov 28 14:33:10 crc kubenswrapper[4817]: I1128 14:33:10.569660 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk7cs" event={"ID":"d0630627-d991-4c3b-ab3b-6a7711ea53f6","Type":"ContainerDied","Data":"b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba"} Nov 28 14:33:10 crc kubenswrapper[4817]: I1128 14:33:10.659111 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:33:10 crc kubenswrapper[4817]: I1128 14:33:10.659146 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:33:10 crc kubenswrapper[4817]: I1128 14:33:10.735776 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.577600 4817 generic.go:334] "Generic (PLEG): container finished" podID="5352e17f-a670-48d4-8ca0-52287875cd21" containerID="5e9ce011b6d5a6bb34c844eb0cb2fdbb74dcb81d574cb61095cc1985c83cad0d" exitCode=0 Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.577673 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c27fw" event={"ID":"5352e17f-a670-48d4-8ca0-52287875cd21","Type":"ContainerDied","Data":"5e9ce011b6d5a6bb34c844eb0cb2fdbb74dcb81d574cb61095cc1985c83cad0d"} Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.580006 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk7cs" event={"ID":"d0630627-d991-4c3b-ab3b-6a7711ea53f6","Type":"ContainerStarted","Data":"cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5"} Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.583146 4817 generic.go:334] "Generic (PLEG): container finished" podID="690548fe-586d-479d-b937-f47a312ef236" containerID="5fe49fc9e0405ca667da6c2483efc9baca3876c0fc8d5d8fcfa0e738dd44909c" exitCode=0 Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.583220 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsnzb" event={"ID":"690548fe-586d-479d-b937-f47a312ef236","Type":"ContainerDied","Data":"5fe49fc9e0405ca667da6c2483efc9baca3876c0fc8d5d8fcfa0e738dd44909c"} Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.585874 4817 generic.go:334] "Generic (PLEG): container finished" podID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerID="8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8" exitCode=0 Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.585935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4l97" event={"ID":"041e51f8-a1ce-41a6-adba-b0b8bd638de3","Type":"ContainerDied","Data":"8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8"} Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.634105 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:33:11 crc kubenswrapper[4817]: I1128 14:33:11.654172 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lk7cs" podStartSLOduration=2.633325827 podStartE2EDuration="1m1.654154587s" podCreationTimestamp="2025-11-28 14:32:10 +0000 UTC" firstStartedPulling="2025-11-28 14:32:12.092967922 +0000 UTC m=+154.680946188" lastFinishedPulling="2025-11-28 14:33:11.113796682 +0000 UTC m=+213.701774948" observedRunningTime="2025-11-28 14:33:11.65156623 +0000 UTC m=+214.239544496" watchObservedRunningTime="2025-11-28 14:33:11.654154587 +0000 UTC m=+214.242132853" Nov 28 14:33:12 crc kubenswrapper[4817]: I1128 14:33:12.592574 4817 generic.go:334] "Generic (PLEG): container finished" podID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerID="cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431" exitCode=0 Nov 28 14:33:12 crc kubenswrapper[4817]: I1128 14:33:12.592759 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p7n" event={"ID":"76973893-1cf3-40bd-aa5e-e2a4a32ec10a","Type":"ContainerDied","Data":"cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431"} Nov 28 14:33:12 crc kubenswrapper[4817]: I1128 14:33:12.597167 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsnzb" event={"ID":"690548fe-586d-479d-b937-f47a312ef236","Type":"ContainerStarted","Data":"0206f5f540fbcbd4fbb5b7d90ee94e21f4fb6e562c8df27002dae895bd0db8fb"} Nov 28 14:33:12 crc kubenswrapper[4817]: I1128 14:33:12.627642 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tsnzb" podStartSLOduration=2.750074306 podStartE2EDuration="59.627624266s" podCreationTimestamp="2025-11-28 14:32:13 +0000 UTC" firstStartedPulling="2025-11-28 14:32:15.193494158 +0000 UTC m=+157.781472424" lastFinishedPulling="2025-11-28 14:33:12.071044098 +0000 UTC m=+214.659022384" observedRunningTime="2025-11-28 14:33:12.625539922 +0000 UTC m=+215.213518188" watchObservedRunningTime="2025-11-28 14:33:12.627624266 +0000 UTC m=+215.215602532" Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.604600 4817 generic.go:334] "Generic (PLEG): container finished" podID="d691474c-0277-427e-abec-03f8df5fa5c1" containerID="5382c3813941bce52fc05318cdaf0ee9b0ada1913da26826924c2572eefabfc2" exitCode=0 Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.604677 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwm9b" event={"ID":"d691474c-0277-427e-abec-03f8df5fa5c1","Type":"ContainerDied","Data":"5382c3813941bce52fc05318cdaf0ee9b0ada1913da26826924c2572eefabfc2"} Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.608638 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerStarted","Data":"11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411"} Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.618249 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p7n" event={"ID":"76973893-1cf3-40bd-aa5e-e2a4a32ec10a","Type":"ContainerStarted","Data":"64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff"} Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.621357 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4l97" event={"ID":"041e51f8-a1ce-41a6-adba-b0b8bd638de3","Type":"ContainerStarted","Data":"49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03"} Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.624878 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c27fw" event={"ID":"5352e17f-a670-48d4-8ca0-52287875cd21","Type":"ContainerStarted","Data":"8bb819eb752934e18e0da24014a113c6de78b54fe8c532c52756161e43767a4c"} Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.634791 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.635558 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.686204 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c27fw" podStartSLOduration=3.507046215 podStartE2EDuration="1m3.686183453s" podCreationTimestamp="2025-11-28 14:32:10 +0000 UTC" firstStartedPulling="2025-11-28 14:32:12.053421093 +0000 UTC m=+154.641399359" lastFinishedPulling="2025-11-28 14:33:12.232558331 +0000 UTC m=+214.820536597" observedRunningTime="2025-11-28 14:33:13.657837118 +0000 UTC m=+216.245815384" watchObservedRunningTime="2025-11-28 14:33:13.686183453 +0000 UTC m=+216.274161719" Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.687169 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c4l97" podStartSLOduration=3.543216583 podStartE2EDuration="1m1.687163819s" podCreationTimestamp="2025-11-28 14:32:12 +0000 UTC" firstStartedPulling="2025-11-28 14:32:14.150257575 +0000 UTC m=+156.738235841" lastFinishedPulling="2025-11-28 14:33:12.294204811 +0000 UTC m=+214.882183077" observedRunningTime="2025-11-28 14:33:13.685618439 +0000 UTC m=+216.273596705" watchObservedRunningTime="2025-11-28 14:33:13.687163819 +0000 UTC m=+216.275142085" Nov 28 14:33:13 crc kubenswrapper[4817]: I1128 14:33:13.726273 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-69p7n" podStartSLOduration=2.758433572 podStartE2EDuration="1m0.726259034s" podCreationTimestamp="2025-11-28 14:32:13 +0000 UTC" firstStartedPulling="2025-11-28 14:32:15.199461062 +0000 UTC m=+157.787439328" lastFinishedPulling="2025-11-28 14:33:13.167286524 +0000 UTC m=+215.755264790" observedRunningTime="2025-11-28 14:33:13.723628385 +0000 UTC m=+216.311606661" watchObservedRunningTime="2025-11-28 14:33:13.726259034 +0000 UTC m=+216.314237300" Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.020424 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.020657 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.631407 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwm9b" event={"ID":"d691474c-0277-427e-abec-03f8df5fa5c1","Type":"ContainerStarted","Data":"1dc3ebabaecd6215233574bf4ef78c0da99051dc5f2f085020f3046362241638"} Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.633638 4817 generic.go:334] "Generic (PLEG): container finished" podID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerID="11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411" exitCode=0 Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.634072 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerDied","Data":"11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411"} Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.649551 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cwm9b" podStartSLOduration=2.677828896 podStartE2EDuration="1m4.64953309s" podCreationTimestamp="2025-11-28 14:32:10 +0000 UTC" firstStartedPulling="2025-11-28 14:32:12.07969823 +0000 UTC m=+154.667676496" lastFinishedPulling="2025-11-28 14:33:14.051402434 +0000 UTC m=+216.639380690" observedRunningTime="2025-11-28 14:33:14.649001886 +0000 UTC m=+217.236980152" watchObservedRunningTime="2025-11-28 14:33:14.64953309 +0000 UTC m=+217.237511356" Nov 28 14:33:14 crc kubenswrapper[4817]: I1128 14:33:14.677117 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-69p7n" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="registry-server" probeResult="failure" output=< Nov 28 14:33:14 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 14:33:14 crc kubenswrapper[4817]: > Nov 28 14:33:15 crc kubenswrapper[4817]: I1128 14:33:15.065240 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tsnzb" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="registry-server" probeResult="failure" output=< Nov 28 14:33:15 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 14:33:15 crc kubenswrapper[4817]: > Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.465358 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.465797 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.522100 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.713080 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.844759 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.844807 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:33:20 crc kubenswrapper[4817]: I1128 14:33:20.912934 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:33:21 crc kubenswrapper[4817]: I1128 14:33:21.057306 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:33:21 crc kubenswrapper[4817]: I1128 14:33:21.057352 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:33:21 crc kubenswrapper[4817]: I1128 14:33:21.102851 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:33:21 crc kubenswrapper[4817]: I1128 14:33:21.719918 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:33:21 crc kubenswrapper[4817]: I1128 14:33:21.750675 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:33:22 crc kubenswrapper[4817]: I1128 14:33:22.628637 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:33:22 crc kubenswrapper[4817]: I1128 14:33:22.629026 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:33:22 crc kubenswrapper[4817]: I1128 14:33:22.666696 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:33:22 crc kubenswrapper[4817]: I1128 14:33:22.709998 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:33:22 crc kubenswrapper[4817]: I1128 14:33:22.778913 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cwm9b"] Nov 28 14:33:23 crc kubenswrapper[4817]: I1128 14:33:23.370475 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c27fw"] Nov 28 14:33:23 crc kubenswrapper[4817]: I1128 14:33:23.679672 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c27fw" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="registry-server" containerID="cri-o://8bb819eb752934e18e0da24014a113c6de78b54fe8c532c52756161e43767a4c" gracePeriod=2 Nov 28 14:33:23 crc kubenswrapper[4817]: I1128 14:33:23.680003 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cwm9b" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="registry-server" containerID="cri-o://1dc3ebabaecd6215233574bf4ef78c0da99051dc5f2f085020f3046362241638" gracePeriod=2 Nov 28 14:33:23 crc kubenswrapper[4817]: I1128 14:33:23.696776 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:33:23 crc kubenswrapper[4817]: I1128 14:33:23.766457 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.045152 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.045229 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.045291 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.046185 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.046341 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a" gracePeriod=600 Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.100605 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.180327 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.685862 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a" exitCode=0 Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.685935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a"} Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.688625 4817 generic.go:334] "Generic (PLEG): container finished" podID="5352e17f-a670-48d4-8ca0-52287875cd21" containerID="8bb819eb752934e18e0da24014a113c6de78b54fe8c532c52756161e43767a4c" exitCode=0 Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.688684 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c27fw" event={"ID":"5352e17f-a670-48d4-8ca0-52287875cd21","Type":"ContainerDied","Data":"8bb819eb752934e18e0da24014a113c6de78b54fe8c532c52756161e43767a4c"} Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.690611 4817 generic.go:334] "Generic (PLEG): container finished" podID="d691474c-0277-427e-abec-03f8df5fa5c1" containerID="1dc3ebabaecd6215233574bf4ef78c0da99051dc5f2f085020f3046362241638" exitCode=0 Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.690660 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwm9b" event={"ID":"d691474c-0277-427e-abec-03f8df5fa5c1","Type":"ContainerDied","Data":"1dc3ebabaecd6215233574bf4ef78c0da99051dc5f2f085020f3046362241638"} Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.692693 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerStarted","Data":"79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532"} Nov 28 14:33:24 crc kubenswrapper[4817]: I1128 14:33:24.721138 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-phfbn" podStartSLOduration=5.470558153 podStartE2EDuration="1m12.72111699s" podCreationTimestamp="2025-11-28 14:32:12 +0000 UTC" firstStartedPulling="2025-11-28 14:32:15.202840799 +0000 UTC m=+157.790819065" lastFinishedPulling="2025-11-28 14:33:22.453399606 +0000 UTC m=+225.041377902" observedRunningTime="2025-11-28 14:33:24.714320024 +0000 UTC m=+227.302298300" watchObservedRunningTime="2025-11-28 14:33:24.72111699 +0000 UTC m=+227.309095256" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.444795 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.494343 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.530123 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-utilities\") pod \"5352e17f-a670-48d4-8ca0-52287875cd21\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.530249 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-catalog-content\") pod \"5352e17f-a670-48d4-8ca0-52287875cd21\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.530301 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh5ms\" (UniqueName: \"kubernetes.io/projected/5352e17f-a670-48d4-8ca0-52287875cd21-kube-api-access-lh5ms\") pod \"5352e17f-a670-48d4-8ca0-52287875cd21\" (UID: \"5352e17f-a670-48d4-8ca0-52287875cd21\") " Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.531321 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-utilities" (OuterVolumeSpecName: "utilities") pod "5352e17f-a670-48d4-8ca0-52287875cd21" (UID: "5352e17f-a670-48d4-8ca0-52287875cd21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.536331 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5352e17f-a670-48d4-8ca0-52287875cd21-kube-api-access-lh5ms" (OuterVolumeSpecName: "kube-api-access-lh5ms") pod "5352e17f-a670-48d4-8ca0-52287875cd21" (UID: "5352e17f-a670-48d4-8ca0-52287875cd21"). InnerVolumeSpecName "kube-api-access-lh5ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.589276 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5352e17f-a670-48d4-8ca0-52287875cd21" (UID: "5352e17f-a670-48d4-8ca0-52287875cd21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.631868 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-utilities\") pod \"d691474c-0277-427e-abec-03f8df5fa5c1\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.632238 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb666\" (UniqueName: \"kubernetes.io/projected/d691474c-0277-427e-abec-03f8df5fa5c1-kube-api-access-xb666\") pod \"d691474c-0277-427e-abec-03f8df5fa5c1\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.632324 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-catalog-content\") pod \"d691474c-0277-427e-abec-03f8df5fa5c1\" (UID: \"d691474c-0277-427e-abec-03f8df5fa5c1\") " Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.632607 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-utilities" (OuterVolumeSpecName: "utilities") pod "d691474c-0277-427e-abec-03f8df5fa5c1" (UID: "d691474c-0277-427e-abec-03f8df5fa5c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.632632 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.632710 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh5ms\" (UniqueName: \"kubernetes.io/projected/5352e17f-a670-48d4-8ca0-52287875cd21-kube-api-access-lh5ms\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.632744 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5352e17f-a670-48d4-8ca0-52287875cd21-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.636850 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d691474c-0277-427e-abec-03f8df5fa5c1-kube-api-access-xb666" (OuterVolumeSpecName: "kube-api-access-xb666") pod "d691474c-0277-427e-abec-03f8df5fa5c1" (UID: "d691474c-0277-427e-abec-03f8df5fa5c1"). InnerVolumeSpecName "kube-api-access-xb666". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.700509 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"7d55fccdcaeded99ec2f598c53c1bf31519abf005e37cbd34c5557b27ee85582"} Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.702300 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c27fw" event={"ID":"5352e17f-a670-48d4-8ca0-52287875cd21","Type":"ContainerDied","Data":"55e1f5c9341f6cbad3141870a7ec9fe2a20bb1b505e90d9a019e437c68b18469"} Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.702366 4817 scope.go:117] "RemoveContainer" containerID="8bb819eb752934e18e0da24014a113c6de78b54fe8c532c52756161e43767a4c" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.702375 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c27fw" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.704484 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cwm9b" event={"ID":"d691474c-0277-427e-abec-03f8df5fa5c1","Type":"ContainerDied","Data":"e391416a92847cc7aa16e6a726c3082b16e3de1f52ed40a85d59198a8bef9b76"} Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.704582 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cwm9b" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.724664 4817 scope.go:117] "RemoveContainer" containerID="5e9ce011b6d5a6bb34c844eb0cb2fdbb74dcb81d574cb61095cc1985c83cad0d" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.734011 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.734049 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb666\" (UniqueName: \"kubernetes.io/projected/d691474c-0277-427e-abec-03f8df5fa5c1-kube-api-access-xb666\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.734993 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d691474c-0277-427e-abec-03f8df5fa5c1" (UID: "d691474c-0277-427e-abec-03f8df5fa5c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.744825 4817 scope.go:117] "RemoveContainer" containerID="e49a9281d168440d3cde0b6c358fa79a433afec381e9076f8f091c36f89952ef" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.756399 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c27fw"] Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.756467 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c27fw"] Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.774121 4817 scope.go:117] "RemoveContainer" containerID="1dc3ebabaecd6215233574bf4ef78c0da99051dc5f2f085020f3046362241638" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.789885 4817 scope.go:117] "RemoveContainer" containerID="5382c3813941bce52fc05318cdaf0ee9b0ada1913da26826924c2572eefabfc2" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.827547 4817 scope.go:117] "RemoveContainer" containerID="30eb8390d1e8ff0d4b3ff57988f63db08dec69b6b5389419e71a4a7dacc1eb1e" Nov 28 14:33:25 crc kubenswrapper[4817]: I1128 14:33:25.835283 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d691474c-0277-427e-abec-03f8df5fa5c1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:26 crc kubenswrapper[4817]: I1128 14:33:26.029730 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cwm9b"] Nov 28 14:33:26 crc kubenswrapper[4817]: I1128 14:33:26.033778 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cwm9b"] Nov 28 14:33:27 crc kubenswrapper[4817]: I1128 14:33:27.742793 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" path="/var/lib/kubelet/pods/5352e17f-a670-48d4-8ca0-52287875cd21/volumes" Nov 28 14:33:27 crc kubenswrapper[4817]: I1128 14:33:27.743974 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" path="/var/lib/kubelet/pods/d691474c-0277-427e-abec-03f8df5fa5c1/volumes" Nov 28 14:33:28 crc kubenswrapper[4817]: I1128 14:33:28.170036 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tsnzb"] Nov 28 14:33:28 crc kubenswrapper[4817]: I1128 14:33:28.170298 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tsnzb" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="registry-server" containerID="cri-o://0206f5f540fbcbd4fbb5b7d90ee94e21f4fb6e562c8df27002dae895bd0db8fb" gracePeriod=2 Nov 28 14:33:28 crc kubenswrapper[4817]: I1128 14:33:28.724950 4817 generic.go:334] "Generic (PLEG): container finished" podID="690548fe-586d-479d-b937-f47a312ef236" containerID="0206f5f540fbcbd4fbb5b7d90ee94e21f4fb6e562c8df27002dae895bd0db8fb" exitCode=0 Nov 28 14:33:28 crc kubenswrapper[4817]: I1128 14:33:28.725064 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsnzb" event={"ID":"690548fe-586d-479d-b937-f47a312ef236","Type":"ContainerDied","Data":"0206f5f540fbcbd4fbb5b7d90ee94e21f4fb6e562c8df27002dae895bd0db8fb"} Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.646405 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.682047 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqc6r\" (UniqueName: \"kubernetes.io/projected/690548fe-586d-479d-b937-f47a312ef236-kube-api-access-tqc6r\") pod \"690548fe-586d-479d-b937-f47a312ef236\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.682115 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-catalog-content\") pod \"690548fe-586d-479d-b937-f47a312ef236\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.682268 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-utilities\") pod \"690548fe-586d-479d-b937-f47a312ef236\" (UID: \"690548fe-586d-479d-b937-f47a312ef236\") " Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.683134 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-utilities" (OuterVolumeSpecName: "utilities") pod "690548fe-586d-479d-b937-f47a312ef236" (UID: "690548fe-586d-479d-b937-f47a312ef236"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.689918 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690548fe-586d-479d-b937-f47a312ef236-kube-api-access-tqc6r" (OuterVolumeSpecName: "kube-api-access-tqc6r") pod "690548fe-586d-479d-b937-f47a312ef236" (UID: "690548fe-586d-479d-b937-f47a312ef236"). InnerVolumeSpecName "kube-api-access-tqc6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.732682 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tsnzb" event={"ID":"690548fe-586d-479d-b937-f47a312ef236","Type":"ContainerDied","Data":"fcaa4a7d476cf8e869ff802aaf3efc1ab4cc9402099d61a7ebebad9fcf2ccded"} Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.732734 4817 scope.go:117] "RemoveContainer" containerID="0206f5f540fbcbd4fbb5b7d90ee94e21f4fb6e562c8df27002dae895bd0db8fb" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.732829 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tsnzb" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.756843 4817 scope.go:117] "RemoveContainer" containerID="5fe49fc9e0405ca667da6c2483efc9baca3876c0fc8d5d8fcfa0e738dd44909c" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.782863 4817 scope.go:117] "RemoveContainer" containerID="2c4b9c980bfc8d9c306e82d4d7ac400ca99629968e2b73ead39bba377d33d5a0" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.783579 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.783608 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqc6r\" (UniqueName: \"kubernetes.io/projected/690548fe-586d-479d-b937-f47a312ef236-kube-api-access-tqc6r\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.815957 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "690548fe-586d-479d-b937-f47a312ef236" (UID: "690548fe-586d-479d-b937-f47a312ef236"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:29 crc kubenswrapper[4817]: I1128 14:33:29.884480 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690548fe-586d-479d-b937-f47a312ef236-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:30 crc kubenswrapper[4817]: I1128 14:33:30.055895 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tsnzb"] Nov 28 14:33:30 crc kubenswrapper[4817]: I1128 14:33:30.062925 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tsnzb"] Nov 28 14:33:31 crc kubenswrapper[4817]: I1128 14:33:31.748542 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690548fe-586d-479d-b937-f47a312ef236" path="/var/lib/kubelet/pods/690548fe-586d-479d-b937-f47a312ef236/volumes" Nov 28 14:33:33 crc kubenswrapper[4817]: I1128 14:33:33.028739 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:33:33 crc kubenswrapper[4817]: I1128 14:33:33.028861 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:33:33 crc kubenswrapper[4817]: I1128 14:33:33.066458 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:33:33 crc kubenswrapper[4817]: I1128 14:33:33.638352 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jczpw"] Nov 28 14:33:33 crc kubenswrapper[4817]: I1128 14:33:33.828908 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.853762 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wp6q4"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.854263 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wp6q4" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="registry-server" containerID="cri-o://c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd" gracePeriod=30 Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.860151 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lk7cs"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.860432 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lk7cs" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="registry-server" containerID="cri-o://cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5" gracePeriod=30 Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.869658 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l6fvp"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.869925 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerName="marketplace-operator" containerID="cri-o://332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8" gracePeriod=30 Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.877842 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4l97"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.878128 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c4l97" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="registry-server" containerID="cri-o://49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03" gracePeriod=30 Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.887951 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.890824 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hxjk5"] Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891100 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891123 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891140 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="extract-utilities" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891150 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="extract-utilities" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891169 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="extract-utilities" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891178 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="extract-utilities" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891187 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="extract-content" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891195 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="extract-content" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891208 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="extract-content" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891215 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="extract-content" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891228 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="extract-utilities" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891236 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="extract-utilities" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891246 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891254 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891263 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="extract-content" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891271 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="extract-content" Nov 28 14:33:34 crc kubenswrapper[4817]: E1128 14:33:34.891285 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891293 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891415 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="690548fe-586d-479d-b937-f47a312ef236" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891435 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5352e17f-a670-48d4-8ca0-52287875cd21" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891446 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d691474c-0277-427e-abec-03f8df5fa5c1" containerName="registry-server" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.891919 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.894401 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69p7n"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.894908 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-69p7n" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="registry-server" containerID="cri-o://64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff" gracePeriod=30 Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.901991 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hxjk5"] Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.944994 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9482d966-5b89-4687-9b4a-29eefc683dcd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.945035 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f2dm\" (UniqueName: \"kubernetes.io/projected/9482d966-5b89-4687-9b4a-29eefc683dcd-kube-api-access-6f2dm\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:34 crc kubenswrapper[4817]: I1128 14:33:34.945065 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9482d966-5b89-4687-9b4a-29eefc683dcd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.047949 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9482d966-5b89-4687-9b4a-29eefc683dcd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.047995 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f2dm\" (UniqueName: \"kubernetes.io/projected/9482d966-5b89-4687-9b4a-29eefc683dcd-kube-api-access-6f2dm\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.048022 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9482d966-5b89-4687-9b4a-29eefc683dcd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.053018 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9482d966-5b89-4687-9b4a-29eefc683dcd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.056050 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9482d966-5b89-4687-9b4a-29eefc683dcd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.069503 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f2dm\" (UniqueName: \"kubernetes.io/projected/9482d966-5b89-4687-9b4a-29eefc683dcd-kube-api-access-6f2dm\") pod \"marketplace-operator-79b997595-hxjk5\" (UID: \"9482d966-5b89-4687-9b4a-29eefc683dcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.227052 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.370011 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.395235 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.396195 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.459277 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4gpl\" (UniqueName: \"kubernetes.io/projected/d0630627-d991-4c3b-ab3b-6a7711ea53f6-kube-api-access-q4gpl\") pod \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.459788 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-catalog-content\") pod \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.459833 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-utilities\") pod \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.459873 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sdcb\" (UniqueName: \"kubernetes.io/projected/041e51f8-a1ce-41a6-adba-b0b8bd638de3-kube-api-access-9sdcb\") pod \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.459914 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-catalog-content\") pod \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\" (UID: \"d0630627-d991-4c3b-ab3b-6a7711ea53f6\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.460000 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-trusted-ca\") pod \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.460045 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-operator-metrics\") pod \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.460127 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2fqh\" (UniqueName: \"kubernetes.io/projected/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-kube-api-access-n2fqh\") pod \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\" (UID: \"26e26dfe-f9df-4fff-8eee-f8895b76e4a4\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.460181 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-utilities\") pod \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\" (UID: \"041e51f8-a1ce-41a6-adba-b0b8bd638de3\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.461248 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-utilities" (OuterVolumeSpecName: "utilities") pod "d0630627-d991-4c3b-ab3b-6a7711ea53f6" (UID: "d0630627-d991-4c3b-ab3b-6a7711ea53f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.461728 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "26e26dfe-f9df-4fff-8eee-f8895b76e4a4" (UID: "26e26dfe-f9df-4fff-8eee-f8895b76e4a4"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.463175 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-utilities" (OuterVolumeSpecName: "utilities") pod "041e51f8-a1ce-41a6-adba-b0b8bd638de3" (UID: "041e51f8-a1ce-41a6-adba-b0b8bd638de3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.465616 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0630627-d991-4c3b-ab3b-6a7711ea53f6-kube-api-access-q4gpl" (OuterVolumeSpecName: "kube-api-access-q4gpl") pod "d0630627-d991-4c3b-ab3b-6a7711ea53f6" (UID: "d0630627-d991-4c3b-ab3b-6a7711ea53f6"). InnerVolumeSpecName "kube-api-access-q4gpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.465663 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041e51f8-a1ce-41a6-adba-b0b8bd638de3-kube-api-access-9sdcb" (OuterVolumeSpecName: "kube-api-access-9sdcb") pod "041e51f8-a1ce-41a6-adba-b0b8bd638de3" (UID: "041e51f8-a1ce-41a6-adba-b0b8bd638de3"). InnerVolumeSpecName "kube-api-access-9sdcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.465680 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-kube-api-access-n2fqh" (OuterVolumeSpecName: "kube-api-access-n2fqh") pod "26e26dfe-f9df-4fff-8eee-f8895b76e4a4" (UID: "26e26dfe-f9df-4fff-8eee-f8895b76e4a4"). InnerVolumeSpecName "kube-api-access-n2fqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.477203 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.484858 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "041e51f8-a1ce-41a6-adba-b0b8bd638de3" (UID: "041e51f8-a1ce-41a6-adba-b0b8bd638de3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.503908 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "26e26dfe-f9df-4fff-8eee-f8895b76e4a4" (UID: "26e26dfe-f9df-4fff-8eee-f8895b76e4a4"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.522564 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0630627-d991-4c3b-ab3b-6a7711ea53f6" (UID: "d0630627-d991-4c3b-ab3b-6a7711ea53f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.562622 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-utilities\") pod \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.562747 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkqdt\" (UniqueName: \"kubernetes.io/projected/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-kube-api-access-mkqdt\") pod \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.562791 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-catalog-content\") pod \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\" (UID: \"76973893-1cf3-40bd-aa5e-e2a4a32ec10a\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563078 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563091 4817 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563102 4817 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563112 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2fqh\" (UniqueName: \"kubernetes.io/projected/26e26dfe-f9df-4fff-8eee-f8895b76e4a4-kube-api-access-n2fqh\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563158 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563169 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4gpl\" (UniqueName: \"kubernetes.io/projected/d0630627-d991-4c3b-ab3b-6a7711ea53f6-kube-api-access-q4gpl\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563178 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/041e51f8-a1ce-41a6-adba-b0b8bd638de3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563185 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0630627-d991-4c3b-ab3b-6a7711ea53f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563194 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sdcb\" (UniqueName: \"kubernetes.io/projected/041e51f8-a1ce-41a6-adba-b0b8bd638de3-kube-api-access-9sdcb\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.563366 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-utilities" (OuterVolumeSpecName: "utilities") pod "76973893-1cf3-40bd-aa5e-e2a4a32ec10a" (UID: "76973893-1cf3-40bd-aa5e-e2a4a32ec10a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.567313 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-kube-api-access-mkqdt" (OuterVolumeSpecName: "kube-api-access-mkqdt") pod "76973893-1cf3-40bd-aa5e-e2a4a32ec10a" (UID: "76973893-1cf3-40bd-aa5e-e2a4a32ec10a"). InnerVolumeSpecName "kube-api-access-mkqdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.638973 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.663956 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwdlp\" (UniqueName: \"kubernetes.io/projected/8789692c-b773-4938-8669-5f76d7017c14-kube-api-access-cwdlp\") pod \"8789692c-b773-4938-8669-5f76d7017c14\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.663992 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-catalog-content\") pod \"8789692c-b773-4938-8669-5f76d7017c14\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.664055 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-utilities\") pod \"8789692c-b773-4938-8669-5f76d7017c14\" (UID: \"8789692c-b773-4938-8669-5f76d7017c14\") " Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.664310 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkqdt\" (UniqueName: \"kubernetes.io/projected/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-kube-api-access-mkqdt\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.664323 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.665455 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-utilities" (OuterVolumeSpecName: "utilities") pod "8789692c-b773-4938-8669-5f76d7017c14" (UID: "8789692c-b773-4938-8669-5f76d7017c14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.666712 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8789692c-b773-4938-8669-5f76d7017c14-kube-api-access-cwdlp" (OuterVolumeSpecName: "kube-api-access-cwdlp") pod "8789692c-b773-4938-8669-5f76d7017c14" (UID: "8789692c-b773-4938-8669-5f76d7017c14"). InnerVolumeSpecName "kube-api-access-cwdlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.679095 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76973893-1cf3-40bd-aa5e-e2a4a32ec10a" (UID: "76973893-1cf3-40bd-aa5e-e2a4a32ec10a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.702571 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hxjk5"] Nov 28 14:33:35 crc kubenswrapper[4817]: W1128 14:33:35.708465 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9482d966_5b89_4687_9b4a_29eefc683dcd.slice/crio-1d3f8ab4069d0a07961258339b2ecd5c70d80499b4b647c4d0ad48cfc36d3db9 WatchSource:0}: Error finding container 1d3f8ab4069d0a07961258339b2ecd5c70d80499b4b647c4d0ad48cfc36d3db9: Status 404 returned error can't find the container with id 1d3f8ab4069d0a07961258339b2ecd5c70d80499b4b647c4d0ad48cfc36d3db9 Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.735977 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8789692c-b773-4938-8669-5f76d7017c14" (UID: "8789692c-b773-4938-8669-5f76d7017c14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.766744 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76973893-1cf3-40bd-aa5e-e2a4a32ec10a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.766770 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.766782 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwdlp\" (UniqueName: \"kubernetes.io/projected/8789692c-b773-4938-8669-5f76d7017c14-kube-api-access-cwdlp\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.766794 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8789692c-b773-4938-8669-5f76d7017c14-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.801264 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" event={"ID":"9482d966-5b89-4687-9b4a-29eefc683dcd","Type":"ContainerStarted","Data":"1d3f8ab4069d0a07961258339b2ecd5c70d80499b4b647c4d0ad48cfc36d3db9"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.803382 4817 generic.go:334] "Generic (PLEG): container finished" podID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerID="64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff" exitCode=0 Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.803447 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p7n" event={"ID":"76973893-1cf3-40bd-aa5e-e2a4a32ec10a","Type":"ContainerDied","Data":"64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.803477 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69p7n" event={"ID":"76973893-1cf3-40bd-aa5e-e2a4a32ec10a","Type":"ContainerDied","Data":"ff05589226b87cbe3b31a49d34f9196a00ab2c3e440bf3aaa93437108a309050"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.803492 4817 scope.go:117] "RemoveContainer" containerID="64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.803608 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69p7n" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.806122 4817 generic.go:334] "Generic (PLEG): container finished" podID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerID="332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8" exitCode=0 Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.806184 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" event={"ID":"26e26dfe-f9df-4fff-8eee-f8895b76e4a4","Type":"ContainerDied","Data":"332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.806213 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" event={"ID":"26e26dfe-f9df-4fff-8eee-f8895b76e4a4","Type":"ContainerDied","Data":"a3b9a8e2e253ab5889f9e369f7cfba5997c6ad4642decef32f71b1b97fb29b64"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.806216 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l6fvp" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.811965 4817 generic.go:334] "Generic (PLEG): container finished" podID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerID="49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03" exitCode=0 Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.812015 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4l97" event={"ID":"041e51f8-a1ce-41a6-adba-b0b8bd638de3","Type":"ContainerDied","Data":"49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.812035 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4l97" event={"ID":"041e51f8-a1ce-41a6-adba-b0b8bd638de3","Type":"ContainerDied","Data":"3873eb7585c9fe8d98c6ddaa92b9e68292e644e6ca134c02300aab45c6796f0f"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.812092 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4l97" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.814338 4817 generic.go:334] "Generic (PLEG): container finished" podID="8789692c-b773-4938-8669-5f76d7017c14" containerID="c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd" exitCode=0 Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.814381 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp6q4" event={"ID":"8789692c-b773-4938-8669-5f76d7017c14","Type":"ContainerDied","Data":"c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.814400 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wp6q4" event={"ID":"8789692c-b773-4938-8669-5f76d7017c14","Type":"ContainerDied","Data":"c2039329da8d204f428430342348e1867f2ff8fbd8be6feb61e1681878d3280a"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.814445 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wp6q4" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.817338 4817 generic.go:334] "Generic (PLEG): container finished" podID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerID="cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5" exitCode=0 Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.817407 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lk7cs" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.817431 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk7cs" event={"ID":"d0630627-d991-4c3b-ab3b-6a7711ea53f6","Type":"ContainerDied","Data":"cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.817467 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lk7cs" event={"ID":"d0630627-d991-4c3b-ab3b-6a7711ea53f6","Type":"ContainerDied","Data":"80c570d47874d212a5df983d831c9959a2da9ea00eafdcf605391e85379a4eb8"} Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.833990 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69p7n"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.840843 4817 scope.go:117] "RemoveContainer" containerID="cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.849377 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-69p7n"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.860862 4817 scope.go:117] "RemoveContainer" containerID="e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.861654 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l6fvp"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.865573 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l6fvp"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.870664 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4l97"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.874687 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4l97"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.876734 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wp6q4"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.880317 4817 scope.go:117] "RemoveContainer" containerID="64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.880341 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wp6q4"] Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.883123 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff\": container with ID starting with 64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff not found: ID does not exist" containerID="64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.883167 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff"} err="failed to get container status \"64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff\": rpc error: code = NotFound desc = could not find container \"64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff\": container with ID starting with 64aa807a8b0d3c750bceb96e35c5b3492fec42b56308400ab7e7912efd975cff not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.883191 4817 scope.go:117] "RemoveContainer" containerID="cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.883457 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431\": container with ID starting with cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431 not found: ID does not exist" containerID="cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.883478 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431"} err="failed to get container status \"cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431\": rpc error: code = NotFound desc = could not find container \"cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431\": container with ID starting with cc676396facbbddd6ec549d57e32b34bcb25c7858ed2f006b7312de2fb852431 not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.883490 4817 scope.go:117] "RemoveContainer" containerID="e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.883674 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc\": container with ID starting with e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc not found: ID does not exist" containerID="e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.883692 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc"} err="failed to get container status \"e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc\": rpc error: code = NotFound desc = could not find container \"e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc\": container with ID starting with e08b6bc88a5927e506ca814c027e025dc5a0a5a833b29f70a07c6d0b6e5527fc not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.883704 4817 scope.go:117] "RemoveContainer" containerID="332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.887532 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lk7cs"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.889598 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lk7cs"] Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.896630 4817 scope.go:117] "RemoveContainer" containerID="332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.897000 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8\": container with ID starting with 332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8 not found: ID does not exist" containerID="332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.897041 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8"} err="failed to get container status \"332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8\": rpc error: code = NotFound desc = could not find container \"332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8\": container with ID starting with 332a1fb550bd59c1ab69ffef782c266ac9f0950a81a7c256aed345c08fde0ed8 not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.897068 4817 scope.go:117] "RemoveContainer" containerID="49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.909290 4817 scope.go:117] "RemoveContainer" containerID="8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.920323 4817 scope.go:117] "RemoveContainer" containerID="79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.937785 4817 scope.go:117] "RemoveContainer" containerID="49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.938149 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03\": container with ID starting with 49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03 not found: ID does not exist" containerID="49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.938182 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03"} err="failed to get container status \"49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03\": rpc error: code = NotFound desc = could not find container \"49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03\": container with ID starting with 49275eee47343800b7c883bd1be38c1a95828773a7d55284acdfae52e14b0c03 not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.938209 4817 scope.go:117] "RemoveContainer" containerID="8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.938446 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8\": container with ID starting with 8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8 not found: ID does not exist" containerID="8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.938470 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8"} err="failed to get container status \"8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8\": rpc error: code = NotFound desc = could not find container \"8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8\": container with ID starting with 8ee11090f9742ad196d64c79056bbf864d665840969bc8b5c9630d8938ee59e8 not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.938484 4817 scope.go:117] "RemoveContainer" containerID="79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.938657 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731\": container with ID starting with 79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731 not found: ID does not exist" containerID="79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.938678 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731"} err="failed to get container status \"79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731\": rpc error: code = NotFound desc = could not find container \"79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731\": container with ID starting with 79460570d0f2a6e166a8c7c554863822d6ee3b2b0196928fe2eab1be90bc8731 not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.938691 4817 scope.go:117] "RemoveContainer" containerID="c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.951240 4817 scope.go:117] "RemoveContainer" containerID="4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.969978 4817 scope.go:117] "RemoveContainer" containerID="8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.984257 4817 scope.go:117] "RemoveContainer" containerID="c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.984823 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd\": container with ID starting with c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd not found: ID does not exist" containerID="c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.984856 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd"} err="failed to get container status \"c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd\": rpc error: code = NotFound desc = could not find container \"c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd\": container with ID starting with c9985c508414b3107e4d5ebc7c06d3e0d598a03df1a49703331c076b5a3064cd not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.984879 4817 scope.go:117] "RemoveContainer" containerID="4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.985321 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a\": container with ID starting with 4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a not found: ID does not exist" containerID="4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.985367 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a"} err="failed to get container status \"4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a\": rpc error: code = NotFound desc = could not find container \"4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a\": container with ID starting with 4f9bfa94d0ed684b0f0202ffc83f0cd600ae6714a858beb24e8d513c6eb7318a not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.985396 4817 scope.go:117] "RemoveContainer" containerID="8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76" Nov 28 14:33:35 crc kubenswrapper[4817]: E1128 14:33:35.985741 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76\": container with ID starting with 8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76 not found: ID does not exist" containerID="8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.985766 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76"} err="failed to get container status \"8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76\": rpc error: code = NotFound desc = could not find container \"8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76\": container with ID starting with 8ffd5961424890a39aca945ac6bc5a972f9c483df1b9497b2a50a2af9d3aae76 not found: ID does not exist" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.985780 4817 scope.go:117] "RemoveContainer" containerID="cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5" Nov 28 14:33:35 crc kubenswrapper[4817]: I1128 14:33:35.998117 4817 scope.go:117] "RemoveContainer" containerID="b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.009691 4817 scope.go:117] "RemoveContainer" containerID="989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.058450 4817 scope.go:117] "RemoveContainer" containerID="cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.059112 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5\": container with ID starting with cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5 not found: ID does not exist" containerID="cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.059168 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5"} err="failed to get container status \"cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5\": rpc error: code = NotFound desc = could not find container \"cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5\": container with ID starting with cb506001c4a067fd04adcdf75c0cd2924282c700b2f0e9c19340f124ad769cd5 not found: ID does not exist" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.059204 4817 scope.go:117] "RemoveContainer" containerID="b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.059522 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba\": container with ID starting with b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba not found: ID does not exist" containerID="b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.059565 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba"} err="failed to get container status \"b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba\": rpc error: code = NotFound desc = could not find container \"b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba\": container with ID starting with b330e11d48febe56d8c856a350ae7671bbecfc10ce9eed16ebb5d2f9c5ca61ba not found: ID does not exist" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.059594 4817 scope.go:117] "RemoveContainer" containerID="989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.060156 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3\": container with ID starting with 989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3 not found: ID does not exist" containerID="989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.060182 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3"} err="failed to get container status \"989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3\": rpc error: code = NotFound desc = could not find container \"989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3\": container with ID starting with 989024ba64824e3981159c7eec7c73af2a14b0cc4b3b335b565806060ccd8cc3 not found: ID does not exist" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.586815 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h8h9j"] Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587241 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587279 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587310 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587329 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587361 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587464 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587491 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587507 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587537 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587556 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587582 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587598 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587617 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerName="marketplace-operator" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587632 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerName="marketplace-operator" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587648 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587664 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587688 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587704 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587759 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587776 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587796 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587812 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="extract-utilities" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587836 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587852 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: E1128 14:33:36.587877 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.587891 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="extract-content" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.588115 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.588146 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.588172 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8789692c-b773-4938-8669-5f76d7017c14" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.588191 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" containerName="registry-server" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.588211 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" containerName="marketplace-operator" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.593474 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8h9j"] Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.593625 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.596566 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.679947 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp99q\" (UniqueName: \"kubernetes.io/projected/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-kube-api-access-pp99q\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.680025 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-catalog-content\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.680058 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-utilities\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.781638 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-catalog-content\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.781753 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-utilities\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.781820 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp99q\" (UniqueName: \"kubernetes.io/projected/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-kube-api-access-pp99q\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.782159 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-catalog-content\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.782498 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-utilities\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.818492 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp99q\" (UniqueName: \"kubernetes.io/projected/0c4cbeb0-f40f-4b47-9e11-21878c2f0848-kube-api-access-pp99q\") pod \"redhat-operators-h8h9j\" (UID: \"0c4cbeb0-f40f-4b47-9e11-21878c2f0848\") " pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.827656 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" event={"ID":"9482d966-5b89-4687-9b4a-29eefc683dcd","Type":"ContainerStarted","Data":"f3d850c867fda0d44cdf036be2dde6baa168efd3c992f762ab50aad05b521d12"} Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.828352 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.836000 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.837492 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-phfbn" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="registry-server" containerID="cri-o://79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532" gracePeriod=30 Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.871402 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hxjk5" podStartSLOduration=2.871382287 podStartE2EDuration="2.871382287s" podCreationTimestamp="2025-11-28 14:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:33:36.855013073 +0000 UTC m=+239.442991419" watchObservedRunningTime="2025-11-28 14:33:36.871382287 +0000 UTC m=+239.459360563" Nov 28 14:33:36 crc kubenswrapper[4817]: I1128 14:33:36.907102 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.175025 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q2tpd"] Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.180488 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.181657 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2tpd"] Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.288850 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-utilities\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.288889 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcdhg\" (UniqueName: \"kubernetes.io/projected/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-kube-api-access-fcdhg\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.288992 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-catalog-content\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.313205 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.364346 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8h9j"] Nov 28 14:33:37 crc kubenswrapper[4817]: W1128 14:33:37.370097 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c4cbeb0_f40f_4b47_9e11_21878c2f0848.slice/crio-da65b676d08139c0e3f5ecffd3c936afb6227199817b8a94145cd6b523f7bb49 WatchSource:0}: Error finding container da65b676d08139c0e3f5ecffd3c936afb6227199817b8a94145cd6b523f7bb49: Status 404 returned error can't find the container with id da65b676d08139c0e3f5ecffd3c936afb6227199817b8a94145cd6b523f7bb49 Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.389795 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-catalog-content\") pod \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.389951 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-utilities\") pod \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.390078 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svf2f\" (UniqueName: \"kubernetes.io/projected/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-kube-api-access-svf2f\") pod \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\" (UID: \"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c\") " Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.390632 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-utilities" (OuterVolumeSpecName: "utilities") pod "5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" (UID: "5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.391197 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-utilities\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.391232 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcdhg\" (UniqueName: \"kubernetes.io/projected/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-kube-api-access-fcdhg\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.391286 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-catalog-content\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.391327 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.391536 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-utilities\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.391865 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-catalog-content\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.394517 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-kube-api-access-svf2f" (OuterVolumeSpecName: "kube-api-access-svf2f") pod "5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" (UID: "5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c"). InnerVolumeSpecName "kube-api-access-svf2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.407604 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcdhg\" (UniqueName: \"kubernetes.io/projected/0ef612d5-ca83-442b-bf1c-6f4e16d10be2-kube-api-access-fcdhg\") pod \"redhat-marketplace-q2tpd\" (UID: \"0ef612d5-ca83-442b-bf1c-6f4e16d10be2\") " pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.408663 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" (UID: "5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.492209 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svf2f\" (UniqueName: \"kubernetes.io/projected/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-kube-api-access-svf2f\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.492253 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.502217 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.577979 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.682317 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2tpd"] Nov 28 14:33:37 crc kubenswrapper[4817]: W1128 14:33:37.686948 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ef612d5_ca83_442b_bf1c_6f4e16d10be2.slice/crio-ef58f5f202f4e99bf4b5f5073d0d9ad5c3b6786b6bb2ec53f76ef574f9d03278 WatchSource:0}: Error finding container ef58f5f202f4e99bf4b5f5073d0d9ad5c3b6786b6bb2ec53f76ef574f9d03278: Status 404 returned error can't find the container with id ef58f5f202f4e99bf4b5f5073d0d9ad5c3b6786b6bb2ec53f76ef574f9d03278 Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.751195 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041e51f8-a1ce-41a6-adba-b0b8bd638de3" path="/var/lib/kubelet/pods/041e51f8-a1ce-41a6-adba-b0b8bd638de3/volumes" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.752024 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e26dfe-f9df-4fff-8eee-f8895b76e4a4" path="/var/lib/kubelet/pods/26e26dfe-f9df-4fff-8eee-f8895b76e4a4/volumes" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.752457 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76973893-1cf3-40bd-aa5e-e2a4a32ec10a" path="/var/lib/kubelet/pods/76973893-1cf3-40bd-aa5e-e2a4a32ec10a/volumes" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.753514 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8789692c-b773-4938-8669-5f76d7017c14" path="/var/lib/kubelet/pods/8789692c-b773-4938-8669-5f76d7017c14/volumes" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.754192 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0630627-d991-4c3b-ab3b-6a7711ea53f6" path="/var/lib/kubelet/pods/d0630627-d991-4c3b-ab3b-6a7711ea53f6/volumes" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.842892 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2tpd" event={"ID":"0ef612d5-ca83-442b-bf1c-6f4e16d10be2","Type":"ContainerStarted","Data":"8445171b4840a8a9b1d38561949dee07feb423085df9246743d222ac98e5780a"} Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.842935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2tpd" event={"ID":"0ef612d5-ca83-442b-bf1c-6f4e16d10be2","Type":"ContainerStarted","Data":"ef58f5f202f4e99bf4b5f5073d0d9ad5c3b6786b6bb2ec53f76ef574f9d03278"} Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.844413 4817 generic.go:334] "Generic (PLEG): container finished" podID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" containerID="c3c9b013bd78fb92932eb213f21568438ed055ddadb7f3659284aba580074e14" exitCode=0 Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.844479 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8h9j" event={"ID":"0c4cbeb0-f40f-4b47-9e11-21878c2f0848","Type":"ContainerDied","Data":"c3c9b013bd78fb92932eb213f21568438ed055ddadb7f3659284aba580074e14"} Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.844505 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8h9j" event={"ID":"0c4cbeb0-f40f-4b47-9e11-21878c2f0848","Type":"ContainerStarted","Data":"da65b676d08139c0e3f5ecffd3c936afb6227199817b8a94145cd6b523f7bb49"} Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.847928 4817 generic.go:334] "Generic (PLEG): container finished" podID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerID="79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532" exitCode=0 Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.848029 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phfbn" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.848080 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerDied","Data":"79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532"} Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.848108 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phfbn" event={"ID":"5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c","Type":"ContainerDied","Data":"4b5f6e28647041c823e788c6b6399d6381c81fb66c3483cc66fbb70a6780fbe4"} Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.848127 4817 scope.go:117] "RemoveContainer" containerID="79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.937073 4817 scope.go:117] "RemoveContainer" containerID="11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.966784 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.969579 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-phfbn"] Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.972873 4817 scope.go:117] "RemoveContainer" containerID="9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9" Nov 28 14:33:37 crc kubenswrapper[4817]: I1128 14:33:37.999805 4817 scope.go:117] "RemoveContainer" containerID="79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532" Nov 28 14:33:38 crc kubenswrapper[4817]: E1128 14:33:38.000483 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532\": container with ID starting with 79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532 not found: ID does not exist" containerID="79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.000540 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532"} err="failed to get container status \"79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532\": rpc error: code = NotFound desc = could not find container \"79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532\": container with ID starting with 79d376dd300cf4b7f87058142b6b078fc099d85a510b3fec16dc8d9c69313532 not found: ID does not exist" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.000567 4817 scope.go:117] "RemoveContainer" containerID="11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411" Nov 28 14:33:38 crc kubenswrapper[4817]: E1128 14:33:38.001037 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411\": container with ID starting with 11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411 not found: ID does not exist" containerID="11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.001080 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411"} err="failed to get container status \"11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411\": rpc error: code = NotFound desc = could not find container \"11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411\": container with ID starting with 11377e89179fcd74a36ba2d87bb8770a96ca3a7fd09127274f7043649f8f3411 not found: ID does not exist" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.001096 4817 scope.go:117] "RemoveContainer" containerID="9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9" Nov 28 14:33:38 crc kubenswrapper[4817]: E1128 14:33:38.001498 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9\": container with ID starting with 9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9 not found: ID does not exist" containerID="9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.001532 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9"} err="failed to get container status \"9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9\": rpc error: code = NotFound desc = could not find container \"9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9\": container with ID starting with 9621970982e2d37007375e11c65799d87b83d30ed83a187fd91c8ccba275fdf9 not found: ID does not exist" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.854054 4817 generic.go:334] "Generic (PLEG): container finished" podID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" containerID="8445171b4840a8a9b1d38561949dee07feb423085df9246743d222ac98e5780a" exitCode=0 Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.854132 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2tpd" event={"ID":"0ef612d5-ca83-442b-bf1c-6f4e16d10be2","Type":"ContainerDied","Data":"8445171b4840a8a9b1d38561949dee07feb423085df9246743d222ac98e5780a"} Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.859061 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8h9j" event={"ID":"0c4cbeb0-f40f-4b47-9e11-21878c2f0848","Type":"ContainerStarted","Data":"c5785e37ae8072c7a0f0d521c26eb318f25f4cf2008f48e64589a7f7a9469172"} Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.973797 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d7r7v"] Nov 28 14:33:38 crc kubenswrapper[4817]: E1128 14:33:38.974034 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="extract-utilities" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.974051 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="extract-utilities" Nov 28 14:33:38 crc kubenswrapper[4817]: E1128 14:33:38.974069 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="registry-server" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.974077 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="registry-server" Nov 28 14:33:38 crc kubenswrapper[4817]: E1128 14:33:38.974095 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="extract-content" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.974103 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="extract-content" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.974231 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" containerName="registry-server" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.975609 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.979136 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 14:33:38 crc kubenswrapper[4817]: I1128 14:33:38.983761 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7r7v"] Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.011048 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69b168e-d53d-476a-b6bf-2dbf64d79cca-catalog-content\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.011102 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69b168e-d53d-476a-b6bf-2dbf64d79cca-utilities\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.011161 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqhcj\" (UniqueName: \"kubernetes.io/projected/f69b168e-d53d-476a-b6bf-2dbf64d79cca-kube-api-access-bqhcj\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.112078 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69b168e-d53d-476a-b6bf-2dbf64d79cca-catalog-content\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.112392 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69b168e-d53d-476a-b6bf-2dbf64d79cca-utilities\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.112525 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqhcj\" (UniqueName: \"kubernetes.io/projected/f69b168e-d53d-476a-b6bf-2dbf64d79cca-kube-api-access-bqhcj\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.112632 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69b168e-d53d-476a-b6bf-2dbf64d79cca-catalog-content\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.113147 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69b168e-d53d-476a-b6bf-2dbf64d79cca-utilities\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.134493 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqhcj\" (UniqueName: \"kubernetes.io/projected/f69b168e-d53d-476a-b6bf-2dbf64d79cca-kube-api-access-bqhcj\") pod \"community-operators-d7r7v\" (UID: \"f69b168e-d53d-476a-b6bf-2dbf64d79cca\") " pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.297880 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.513786 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7r7v"] Nov 28 14:33:39 crc kubenswrapper[4817]: W1128 14:33:39.514128 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf69b168e_d53d_476a_b6bf_2dbf64d79cca.slice/crio-a76cccc8a3c27edca9de0f25299d8694519248d272c259ea8e6035aa9ede6449 WatchSource:0}: Error finding container a76cccc8a3c27edca9de0f25299d8694519248d272c259ea8e6035aa9ede6449: Status 404 returned error can't find the container with id a76cccc8a3c27edca9de0f25299d8694519248d272c259ea8e6035aa9ede6449 Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.744939 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c" path="/var/lib/kubelet/pods/5dba559e-aae8-4bb8-8a1f-4b166f5bbf0c/volumes" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.866468 4817 generic.go:334] "Generic (PLEG): container finished" podID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" containerID="17d46532fa804e728e6953b88b27b9b8ded5e3e258e0615125fc7f492d8aa3fe" exitCode=0 Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.866530 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7r7v" event={"ID":"f69b168e-d53d-476a-b6bf-2dbf64d79cca","Type":"ContainerDied","Data":"17d46532fa804e728e6953b88b27b9b8ded5e3e258e0615125fc7f492d8aa3fe"} Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.866594 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7r7v" event={"ID":"f69b168e-d53d-476a-b6bf-2dbf64d79cca","Type":"ContainerStarted","Data":"a76cccc8a3c27edca9de0f25299d8694519248d272c259ea8e6035aa9ede6449"} Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.870505 4817 generic.go:334] "Generic (PLEG): container finished" podID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" containerID="e7ae80ac0e5b011103f6b754bbc6b75ad6d96e06e08369bcc3aaf3962f7739d4" exitCode=0 Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.870645 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2tpd" event={"ID":"0ef612d5-ca83-442b-bf1c-6f4e16d10be2","Type":"ContainerDied","Data":"e7ae80ac0e5b011103f6b754bbc6b75ad6d96e06e08369bcc3aaf3962f7739d4"} Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.873131 4817 generic.go:334] "Generic (PLEG): container finished" podID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" containerID="c5785e37ae8072c7a0f0d521c26eb318f25f4cf2008f48e64589a7f7a9469172" exitCode=0 Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.873170 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8h9j" event={"ID":"0c4cbeb0-f40f-4b47-9e11-21878c2f0848","Type":"ContainerDied","Data":"c5785e37ae8072c7a0f0d521c26eb318f25f4cf2008f48e64589a7f7a9469172"} Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.981113 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r22b9"] Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.983701 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.986204 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 14:33:39 crc kubenswrapper[4817]: I1128 14:33:39.986871 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r22b9"] Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.024495 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzfgm\" (UniqueName: \"kubernetes.io/projected/7e0d724a-6b6d-4fee-8481-7649c214af98-kube-api-access-mzfgm\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.024589 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e0d724a-6b6d-4fee-8481-7649c214af98-catalog-content\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.024660 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e0d724a-6b6d-4fee-8481-7649c214af98-utilities\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.125362 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzfgm\" (UniqueName: \"kubernetes.io/projected/7e0d724a-6b6d-4fee-8481-7649c214af98-kube-api-access-mzfgm\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.125419 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e0d724a-6b6d-4fee-8481-7649c214af98-catalog-content\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.125444 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e0d724a-6b6d-4fee-8481-7649c214af98-utilities\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.125879 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e0d724a-6b6d-4fee-8481-7649c214af98-utilities\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.125952 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e0d724a-6b6d-4fee-8481-7649c214af98-catalog-content\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.146849 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzfgm\" (UniqueName: \"kubernetes.io/projected/7e0d724a-6b6d-4fee-8481-7649c214af98-kube-api-access-mzfgm\") pod \"certified-operators-r22b9\" (UID: \"7e0d724a-6b6d-4fee-8481-7649c214af98\") " pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.305538 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.723536 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r22b9"] Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.877833 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r22b9" event={"ID":"7e0d724a-6b6d-4fee-8481-7649c214af98","Type":"ContainerStarted","Data":"c562f77d4886f7a0ee7ca366b65e17cfb02a54c7b699bc8ef0b989ee3aad4e43"} Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.877873 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r22b9" event={"ID":"7e0d724a-6b6d-4fee-8481-7649c214af98","Type":"ContainerStarted","Data":"ae380e8b7d99a650940973c3870d9be1f2a8215fc42f76d0e0ab3166fe70028a"} Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.880525 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2tpd" event={"ID":"0ef612d5-ca83-442b-bf1c-6f4e16d10be2","Type":"ContainerStarted","Data":"0cffb2306a211a8bfa6d07025f9c60a3da6da5909595517d24aa5893eb944a58"} Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.882497 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8h9j" event={"ID":"0c4cbeb0-f40f-4b47-9e11-21878c2f0848","Type":"ContainerStarted","Data":"7aa2c3e24e35f17e0d98e3a2ce28784ac4c46d54832c79d48ed7fe0047943c96"} Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.884388 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7r7v" event={"ID":"f69b168e-d53d-476a-b6bf-2dbf64d79cca","Type":"ContainerStarted","Data":"714468d5631023604905b178c041b2db54db4c4cada131d6dcbc356c7569d4ed"} Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.908918 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q2tpd" podStartSLOduration=2.447010115 podStartE2EDuration="3.908902371s" podCreationTimestamp="2025-11-28 14:33:37 +0000 UTC" firstStartedPulling="2025-11-28 14:33:38.85549191 +0000 UTC m=+241.443470176" lastFinishedPulling="2025-11-28 14:33:40.317384156 +0000 UTC m=+242.905362432" observedRunningTime="2025-11-28 14:33:40.907695819 +0000 UTC m=+243.495674085" watchObservedRunningTime="2025-11-28 14:33:40.908902371 +0000 UTC m=+243.496880637" Nov 28 14:33:40 crc kubenswrapper[4817]: I1128 14:33:40.946971 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h8h9j" podStartSLOduration=2.441123104 podStartE2EDuration="4.946951358s" podCreationTimestamp="2025-11-28 14:33:36 +0000 UTC" firstStartedPulling="2025-11-28 14:33:37.846959711 +0000 UTC m=+240.434937987" lastFinishedPulling="2025-11-28 14:33:40.352787965 +0000 UTC m=+242.940766241" observedRunningTime="2025-11-28 14:33:40.944162576 +0000 UTC m=+243.532140842" watchObservedRunningTime="2025-11-28 14:33:40.946951358 +0000 UTC m=+243.534929624" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.636762 4817 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.637334 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87" gracePeriod=15 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.637487 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9" gracePeriod=15 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.637526 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d" gracePeriod=15 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.637558 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9" gracePeriod=15 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.637583 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd" gracePeriod=15 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639458 4817 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639785 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639810 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639832 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639844 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639870 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639882 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639897 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639910 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639927 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639939 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639957 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.639969 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.639987 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640000 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640161 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640178 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640193 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640209 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640229 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.640552 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.647347 4817 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.648108 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.653062 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.683361 4817 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.757349 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.757458 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.757518 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.757577 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.757637 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.757702 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.759487 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.759532 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.859930 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860003 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860039 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860042 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860084 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860103 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860138 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860170 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860163 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860199 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860232 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860298 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860396 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860442 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860495 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.860524 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.893699 4817 generic.go:334] "Generic (PLEG): container finished" podID="7e0d724a-6b6d-4fee-8481-7649c214af98" containerID="c562f77d4886f7a0ee7ca366b65e17cfb02a54c7b699bc8ef0b989ee3aad4e43" exitCode=0 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.893830 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r22b9" event={"ID":"7e0d724a-6b6d-4fee-8481-7649c214af98","Type":"ContainerDied","Data":"c562f77d4886f7a0ee7ca366b65e17cfb02a54c7b699bc8ef0b989ee3aad4e43"} Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.894352 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:41 crc kubenswrapper[4817]: E1128 14:33:41.895555 4817 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-r22b9.187c32440ded3ae4 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-r22b9,UID:7e0d724a-6b6d-4fee-8481-7649c214af98,APIVersion:v1,ResourceVersion:29499,FieldPath:spec.initContainers{extract-content},},Reason:Pulling,Message:Pulling image \"registry.redhat.io/redhat/certified-operator-index:v4.18\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 14:33:41.895207652 +0000 UTC m=+244.483185918,LastTimestamp:2025-11-28 14:33:41.895207652 +0000 UTC m=+244.483185918,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.896345 4817 generic.go:334] "Generic (PLEG): container finished" podID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" containerID="714468d5631023604905b178c041b2db54db4c4cada131d6dcbc356c7569d4ed" exitCode=0 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.896416 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7r7v" event={"ID":"f69b168e-d53d-476a-b6bf-2dbf64d79cca","Type":"ContainerDied","Data":"714468d5631023604905b178c041b2db54db4c4cada131d6dcbc356c7569d4ed"} Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.896869 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.897122 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.900504 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.901864 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.902589 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9" exitCode=0 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.902616 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d" exitCode=0 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.902624 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9" exitCode=0 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.902631 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd" exitCode=2 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.902695 4817 scope.go:117] "RemoveContainer" containerID="dccbb0a5c9c5985106447c18717379d9d411bb245fb1c2a6b3a6558affeb7530" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.904507 4817 generic.go:334] "Generic (PLEG): container finished" podID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" containerID="c8a83dbb5b959e70abbae07120c957a37a7bd5c4d96cc2ef314a5ba1c9fce26b" exitCode=0 Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.905159 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"50ce10d4-a7fa-491f-bbe2-50117f74f86c","Type":"ContainerDied","Data":"c8a83dbb5b959e70abbae07120c957a37a7bd5c4d96cc2ef314a5ba1c9fce26b"} Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.906131 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.906345 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.906545 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:41 crc kubenswrapper[4817]: I1128 14:33:41.984054 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:41 crc kubenswrapper[4817]: W1128 14:33:41.999465 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-edd7aad551220ef48722a4270b979d164e4d14c3c87ad818fb6fa4bd5b06023f WatchSource:0}: Error finding container edd7aad551220ef48722a4270b979d164e4d14c3c87ad818fb6fa4bd5b06023f: Status 404 returned error can't find the container with id edd7aad551220ef48722a4270b979d164e4d14c3c87ad818fb6fa4bd5b06023f Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.912913 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7r7v" event={"ID":"f69b168e-d53d-476a-b6bf-2dbf64d79cca","Type":"ContainerStarted","Data":"46a829a91bdd6c3cb982732fcbe702d79d6fbec2e1e42f728ce360cc6ba2c6b3"} Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.915808 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.916211 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.916487 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.920384 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.923658 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r22b9" event={"ID":"7e0d724a-6b6d-4fee-8481-7649c214af98","Type":"ContainerStarted","Data":"911828ea6e8a4d2f2f37e8a6e839ef9592185f2047b76307877c543f1d7aad31"} Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.924627 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.925070 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.925252 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e24185531a0806b3133ffed5328f254f3f4a2e17304e41d45c7d78d99178ed3f"} Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.925371 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"edd7aad551220ef48722a4270b979d164e4d14c3c87ad818fb6fa4bd5b06023f"} Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.925497 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: E1128 14:33:42.926080 4817 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.926133 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.926369 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:42 crc kubenswrapper[4817]: I1128 14:33:42.926596 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.170481 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.171085 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.171243 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.171496 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.279168 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kube-api-access\") pod \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.279465 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-var-lock\") pod \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.279506 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kubelet-dir\") pod \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\" (UID: \"50ce10d4-a7fa-491f-bbe2-50117f74f86c\") " Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.279643 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-var-lock" (OuterVolumeSpecName: "var-lock") pod "50ce10d4-a7fa-491f-bbe2-50117f74f86c" (UID: "50ce10d4-a7fa-491f-bbe2-50117f74f86c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.279754 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "50ce10d4-a7fa-491f-bbe2-50117f74f86c" (UID: "50ce10d4-a7fa-491f-bbe2-50117f74f86c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.280000 4817 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.280020 4817 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.286091 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "50ce10d4-a7fa-491f-bbe2-50117f74f86c" (UID: "50ce10d4-a7fa-491f-bbe2-50117f74f86c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.381435 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50ce10d4-a7fa-491f-bbe2-50117f74f86c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.720131 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:43Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:43Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:43Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:43Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.720774 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.720998 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.721181 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.721368 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.721406 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.930354 4817 generic.go:334] "Generic (PLEG): container finished" podID="7e0d724a-6b6d-4fee-8481-7649c214af98" containerID="911828ea6e8a4d2f2f37e8a6e839ef9592185f2047b76307877c543f1d7aad31" exitCode=0 Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.930432 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r22b9" event={"ID":"7e0d724a-6b6d-4fee-8481-7649c214af98","Type":"ContainerDied","Data":"911828ea6e8a4d2f2f37e8a6e839ef9592185f2047b76307877c543f1d7aad31"} Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.930977 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.932025 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.932377 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.932471 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"50ce10d4-a7fa-491f-bbe2-50117f74f86c","Type":"ContainerDied","Data":"68e89bcde1d2fb28421124843fb29c26f98328a1c36ab8d3c7229f88ca2fb985"} Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.932498 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68e89bcde1d2fb28421124843fb29c26f98328a1c36ab8d3c7229f88ca2fb985" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.932546 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 14:33:43 crc kubenswrapper[4817]: E1128 14:33:43.933337 4817 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.988815 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.989484 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.989753 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.994238 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.994929 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.995259 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.995419 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.995634 4817 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:43 crc kubenswrapper[4817]: I1128 14:33:43.995976 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.090521 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.090603 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.090679 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.090997 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.091039 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.091040 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.192238 4817 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.192302 4817 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.192313 4817 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.941514 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.942900 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87" exitCode=0 Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.943025 4817 scope.go:117] "RemoveContainer" containerID="5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.943041 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.950743 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r22b9" event={"ID":"7e0d724a-6b6d-4fee-8481-7649c214af98","Type":"ContainerStarted","Data":"d1f0d48869ead9d7736eb50437893e3775eaa372504cae1469d36c721a73139f"} Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.951265 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.951494 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.962792 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.963037 4817 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.970125 4817 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.970505 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.970830 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.971079 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.975351 4817 scope.go:117] "RemoveContainer" containerID="d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d" Nov 28 14:33:44 crc kubenswrapper[4817]: I1128 14:33:44.987564 4817 scope.go:117] "RemoveContainer" containerID="460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.004795 4817 scope.go:117] "RemoveContainer" containerID="2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.021662 4817 scope.go:117] "RemoveContainer" containerID="1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.046622 4817 scope.go:117] "RemoveContainer" containerID="e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.081304 4817 scope.go:117] "RemoveContainer" containerID="5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9" Nov 28 14:33:45 crc kubenswrapper[4817]: E1128 14:33:45.081846 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\": container with ID starting with 5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9 not found: ID does not exist" containerID="5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.081897 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9"} err="failed to get container status \"5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\": rpc error: code = NotFound desc = could not find container \"5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9\": container with ID starting with 5c5224b5f2a6050a9970302c2779b0de8506e23bf5e8a4076fce23f23be85de9 not found: ID does not exist" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.081931 4817 scope.go:117] "RemoveContainer" containerID="d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d" Nov 28 14:33:45 crc kubenswrapper[4817]: E1128 14:33:45.082580 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\": container with ID starting with d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d not found: ID does not exist" containerID="d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.082616 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d"} err="failed to get container status \"d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\": rpc error: code = NotFound desc = could not find container \"d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d\": container with ID starting with d711918d3ff61b31b6de2c17e7769a3fb7e65e365d1b3cfd6ac671a2a44f725d not found: ID does not exist" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.082639 4817 scope.go:117] "RemoveContainer" containerID="460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9" Nov 28 14:33:45 crc kubenswrapper[4817]: E1128 14:33:45.082916 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\": container with ID starting with 460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9 not found: ID does not exist" containerID="460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.082936 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9"} err="failed to get container status \"460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\": rpc error: code = NotFound desc = could not find container \"460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9\": container with ID starting with 460d7c56aaeceb862783a6fb0a9ebd4e988e1bf3848cbcae7b94d07af36f8cc9 not found: ID does not exist" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.082952 4817 scope.go:117] "RemoveContainer" containerID="2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd" Nov 28 14:33:45 crc kubenswrapper[4817]: E1128 14:33:45.083276 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\": container with ID starting with 2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd not found: ID does not exist" containerID="2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.083301 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd"} err="failed to get container status \"2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\": rpc error: code = NotFound desc = could not find container \"2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd\": container with ID starting with 2d0a66dbd6e08950910141f69c990ee3deca41f642d7ae550d0c3584cc3404cd not found: ID does not exist" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.083314 4817 scope.go:117] "RemoveContainer" containerID="1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87" Nov 28 14:33:45 crc kubenswrapper[4817]: E1128 14:33:45.084298 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\": container with ID starting with 1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87 not found: ID does not exist" containerID="1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.084323 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87"} err="failed to get container status \"1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\": rpc error: code = NotFound desc = could not find container \"1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87\": container with ID starting with 1859c9b49909afeddd04f79c7b1f0d79e00c1b246d6add87701a68427742de87 not found: ID does not exist" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.084338 4817 scope.go:117] "RemoveContainer" containerID="e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8" Nov 28 14:33:45 crc kubenswrapper[4817]: E1128 14:33:45.085873 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\": container with ID starting with e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8 not found: ID does not exist" containerID="e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.085914 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8"} err="failed to get container status \"e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\": rpc error: code = NotFound desc = could not find container \"e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8\": container with ID starting with e7d938603a30bfb56d252a71a20f39d2e9d911f1e3d81481360a061b37bb7db8 not found: ID does not exist" Nov 28 14:33:45 crc kubenswrapper[4817]: I1128 14:33:45.745308 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 28 14:33:46 crc kubenswrapper[4817]: E1128 14:33:46.342615 4817 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-r22b9.187c32440ded3ae4 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-r22b9,UID:7e0d724a-6b6d-4fee-8481-7649c214af98,APIVersion:v1,ResourceVersion:29499,FieldPath:spec.initContainers{extract-content},},Reason:Pulling,Message:Pulling image \"registry.redhat.io/redhat/certified-operator-index:v4.18\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 14:33:41.895207652 +0000 UTC m=+244.483185918,LastTimestamp:2025-11-28 14:33:41.895207652 +0000 UTC m=+244.483185918,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.908021 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.908060 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.950086 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.950642 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.950859 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.951051 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.951208 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.995413 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h8h9j" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.995950 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.996178 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.996355 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:46 crc kubenswrapper[4817]: I1128 14:33:46.996588 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.614544 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.614928 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.661186 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.661609 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.661920 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.662331 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.662611 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.662884 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.739994 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.740251 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.740709 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.740981 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:47 crc kubenswrapper[4817]: I1128 14:33:47.741231 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:48 crc kubenswrapper[4817]: I1128 14:33:48.004861 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q2tpd" Nov 28 14:33:48 crc kubenswrapper[4817]: I1128 14:33:48.005397 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:48 crc kubenswrapper[4817]: I1128 14:33:48.005764 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:48 crc kubenswrapper[4817]: I1128 14:33:48.006078 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:48 crc kubenswrapper[4817]: I1128 14:33:48.006333 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:48 crc kubenswrapper[4817]: I1128 14:33:48.006570 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.298419 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.298776 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.338920 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.339430 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.339668 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.339918 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.340146 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:49 crc kubenswrapper[4817]: I1128 14:33:49.340353 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.041026 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d7r7v" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.041811 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.042600 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.043289 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.043861 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.044376 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.185900 4817 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.186854 4817 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.187542 4817 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.188338 4817 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.189007 4817 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.189223 4817 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.189856 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="200ms" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.306696 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.306795 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.345263 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.345865 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.346256 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.346816 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.347155 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: I1128 14:33:50.347440 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.391030 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="400ms" Nov 28 14:33:50 crc kubenswrapper[4817]: E1128 14:33:50.791816 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="800ms" Nov 28 14:33:51 crc kubenswrapper[4817]: I1128 14:33:51.028241 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r22b9" Nov 28 14:33:51 crc kubenswrapper[4817]: I1128 14:33:51.028936 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:51 crc kubenswrapper[4817]: I1128 14:33:51.029425 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:51 crc kubenswrapper[4817]: I1128 14:33:51.029672 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:51 crc kubenswrapper[4817]: I1128 14:33:51.029920 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:51 crc kubenswrapper[4817]: I1128 14:33:51.030241 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:51 crc kubenswrapper[4817]: E1128 14:33:51.592625 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="1.6s" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.198477 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="3.2s" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.825860 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:53Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:53Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:53Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T14:33:53Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.826437 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.826940 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.827325 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.827709 4817 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:53 crc kubenswrapper[4817]: E1128 14:33:53.827755 4817 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 14:33:54 crc kubenswrapper[4817]: I1128 14:33:54.534917 4817 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 28 14:33:54 crc kubenswrapper[4817]: I1128 14:33:54.535009 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.010085 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.010140 4817 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a" exitCode=1 Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.010169 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a"} Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.010621 4817 scope.go:117] "RemoveContainer" containerID="a2168cacfa3e45a48e6f5fe50cbb3ee81c3ac3df02ce92d036b0b39f8de8d36a" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.011940 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.013391 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.013774 4817 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.014212 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.014629 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:55 crc kubenswrapper[4817]: I1128 14:33:55.015170 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.018310 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.018371 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c8f41309beee90cb115397af7efae5650c4e6f534f694a4e697a01eb0193c9f2"} Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.019107 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.019671 4817 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.020283 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.020781 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.021112 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.021571 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: E1128 14:33:56.346208 4817 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-r22b9.187c32440ded3ae4 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-r22b9,UID:7e0d724a-6b6d-4fee-8481-7649c214af98,APIVersion:v1,ResourceVersion:29499,FieldPath:spec.initContainers{extract-content},},Reason:Pulling,Message:Pulling image \"registry.redhat.io/redhat/certified-operator-index:v4.18\",Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 14:33:41.895207652 +0000 UTC m=+244.483185918,LastTimestamp:2025-11-28 14:33:41.895207652 +0000 UTC m=+244.483185918,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 14:33:56 crc kubenswrapper[4817]: E1128 14:33:56.399708 4817 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="6.4s" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.737110 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.739401 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.744896 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.745372 4817 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.745816 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.746166 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.746787 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.754647 4817 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.754674 4817 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:33:56 crc kubenswrapper[4817]: E1128 14:33:56.755048 4817 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:56 crc kubenswrapper[4817]: I1128 14:33:56.755788 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:56 crc kubenswrapper[4817]: W1128 14:33:56.787192 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5c5a7a2babb39bf4ec21a6348ffd02c58bc979c067cc50d3b6e090f88b1443cc WatchSource:0}: Error finding container 5c5a7a2babb39bf4ec21a6348ffd02c58bc979c067cc50d3b6e090f88b1443cc: Status 404 returned error can't find the container with id 5c5a7a2babb39bf4ec21a6348ffd02c58bc979c067cc50d3b6e090f88b1443cc Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.022864 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5c5a7a2babb39bf4ec21a6348ffd02c58bc979c067cc50d3b6e090f88b1443cc"} Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.749162 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.750271 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.753544 4817 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.760195 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.760767 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.761302 4817 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:57 crc kubenswrapper[4817]: I1128 14:33:57.761804 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.031594 4817 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b1611194b8c523e40f7f943081932f8a87fb2276f191f887db3670335fbcd81a" exitCode=0 Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.031693 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b1611194b8c523e40f7f943081932f8a87fb2276f191f887db3670335fbcd81a"} Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.032017 4817 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.032265 4817 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.032525 4817 status_manager.go:851] "Failed to get status for pod" podUID="7e0d724a-6b6d-4fee-8481-7649c214af98" pod="openshift-marketplace/certified-operators-r22b9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r22b9\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: E1128 14:33:58.032709 4817 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.033019 4817 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.033400 4817 status_manager.go:851] "Failed to get status for pod" podUID="0ef612d5-ca83-442b-bf1c-6f4e16d10be2" pod="openshift-marketplace/redhat-marketplace-q2tpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q2tpd\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.034703 4817 status_manager.go:851] "Failed to get status for pod" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.035137 4817 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.035712 4817 status_manager.go:851] "Failed to get status for pod" podUID="f69b168e-d53d-476a-b6bf-2dbf64d79cca" pod="openshift-marketplace/community-operators-d7r7v" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-d7r7v\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.036320 4817 status_manager.go:851] "Failed to get status for pod" podUID="0c4cbeb0-f40f-4b47-9e11-21878c2f0848" pod="openshift-marketplace/redhat-operators-h8h9j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-h8h9j\": dial tcp 38.102.83.236:6443: connect: connection refused" Nov 28 14:33:58 crc kubenswrapper[4817]: I1128 14:33:58.707386 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" podUID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" containerName="oauth-openshift" containerID="cri-o://b40ad03709c2ac77d4278d1720ae5d7537ea4c1692cdb9aafc31af60f08a9802" gracePeriod=15 Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.039736 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"52c1fc31364c021785011c495457c1e27cee9d568c1fe5836b1a77b4e82c6cc6"} Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.042300 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" event={"ID":"f890cb5e-2cff-4904-b9f5-3ecd702e869e","Type":"ContainerDied","Data":"b40ad03709c2ac77d4278d1720ae5d7537ea4c1692cdb9aafc31af60f08a9802"} Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.042376 4817 generic.go:334] "Generic (PLEG): container finished" podID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" containerID="b40ad03709c2ac77d4278d1720ae5d7537ea4c1692cdb9aafc31af60f08a9802" exitCode=0 Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.420665 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577351 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-idp-0-file-data\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577413 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6pwd\" (UniqueName: \"kubernetes.io/projected/f890cb5e-2cff-4904-b9f5-3ecd702e869e-kube-api-access-b6pwd\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577441 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-trusted-ca-bundle\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577494 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-provider-selection\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577520 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-service-ca\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577556 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-session\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577578 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-router-certs\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577606 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-cliconfig\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577652 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-dir\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577679 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-serving-cert\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577705 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-policies\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577780 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-error\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577803 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-login\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.577827 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-ocp-branding-template\") pod \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\" (UID: \"f890cb5e-2cff-4904-b9f5-3ecd702e869e\") " Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.578380 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.578489 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.578532 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.578602 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.587477 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.589993 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.590433 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.590548 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.588323 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.590707 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.590958 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.592774 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.602050 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f890cb5e-2cff-4904-b9f5-3ecd702e869e-kube-api-access-b6pwd" (OuterVolumeSpecName: "kube-api-access-b6pwd") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "kube-api-access-b6pwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.602470 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f890cb5e-2cff-4904-b9f5-3ecd702e869e" (UID: "f890cb5e-2cff-4904-b9f5-3ecd702e869e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678849 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678887 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678902 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678916 4817 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678930 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678942 4817 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678953 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678965 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678977 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678988 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.678998 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.679008 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6pwd\" (UniqueName: \"kubernetes.io/projected/f890cb5e-2cff-4904-b9f5-3ecd702e869e-kube-api-access-b6pwd\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.679020 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 14:33:59 crc kubenswrapper[4817]: I1128 14:33:59.679031 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f890cb5e-2cff-4904-b9f5-3ecd702e869e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:34:00 crc kubenswrapper[4817]: I1128 14:34:00.050219 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fd44491ceb20fe4ec7c5a53a164ae82f298e112a716fcc62aaa9d386db9b9c67"} Nov 28 14:34:00 crc kubenswrapper[4817]: I1128 14:34:00.050270 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e16e6c894b9ef7537ae54c2e50e471337a14777fa015f486382dbf1055b02aa"} Nov 28 14:34:00 crc kubenswrapper[4817]: I1128 14:34:00.051911 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" event={"ID":"f890cb5e-2cff-4904-b9f5-3ecd702e869e","Type":"ContainerDied","Data":"d92ffcbb3226cbf663e02caaa8976ddf0c8638f88585bf793da6c91f8574ae81"} Nov 28 14:34:00 crc kubenswrapper[4817]: I1128 14:34:00.051989 4817 scope.go:117] "RemoveContainer" containerID="b40ad03709c2ac77d4278d1720ae5d7537ea4c1692cdb9aafc31af60f08a9802" Nov 28 14:34:00 crc kubenswrapper[4817]: I1128 14:34:00.051994 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jczpw" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.063754 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d6634380e5373a512b8054bea61c73f13198ca455a191c1f53286a840f64b06b"} Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.064935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2ebc14da88c2a17decebfbe71c3892871f1ef9a377db9be64bb2caa02a6f6975"} Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.065045 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.064450 4817 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.065228 4817 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.755996 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.756481 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.764563 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.935081 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:34:01 crc kubenswrapper[4817]: I1128 14:34:01.942001 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:34:02 crc kubenswrapper[4817]: I1128 14:34:02.071588 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:34:06 crc kubenswrapper[4817]: I1128 14:34:06.076716 4817 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:06 crc kubenswrapper[4817]: I1128 14:34:06.760831 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:06 crc kubenswrapper[4817]: I1128 14:34:06.763094 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="231efe72-bfdb-498c-bda2-045306c2a5b2" Nov 28 14:34:07 crc kubenswrapper[4817]: I1128 14:34:07.105690 4817 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:07 crc kubenswrapper[4817]: I1128 14:34:07.105780 4817 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:07 crc kubenswrapper[4817]: I1128 14:34:07.773411 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="231efe72-bfdb-498c-bda2-045306c2a5b2" Nov 28 14:34:08 crc kubenswrapper[4817]: I1128 14:34:08.113348 4817 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:08 crc kubenswrapper[4817]: I1128 14:34:08.113928 4817 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:08 crc kubenswrapper[4817]: I1128 14:34:08.121152 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="231efe72-bfdb-498c-bda2-045306c2a5b2" Nov 28 14:34:13 crc kubenswrapper[4817]: I1128 14:34:13.666016 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.410393 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.555671 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.585674 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.648301 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.702122 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.837486 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 14:34:15 crc kubenswrapper[4817]: I1128 14:34:15.976864 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 14:34:16 crc kubenswrapper[4817]: I1128 14:34:16.028191 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 14:34:16 crc kubenswrapper[4817]: I1128 14:34:16.307591 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 14:34:16 crc kubenswrapper[4817]: I1128 14:34:16.315167 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 14:34:16 crc kubenswrapper[4817]: I1128 14:34:16.687455 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:16.722982 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:16.793677 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:16.842939 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:17.004489 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:17.393525 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:17.394661 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:17.398060 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:17.582021 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 14:34:17 crc kubenswrapper[4817]: I1128 14:34:17.743079 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.019268 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.073192 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.079982 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.090362 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.101796 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.122786 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.157084 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.189121 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.210774 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.334388 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.353547 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.359019 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.481050 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.666467 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.872245 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 14:34:18 crc kubenswrapper[4817]: I1128 14:34:18.898168 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.017242 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.184559 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.211197 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.460335 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.602116 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.637566 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.745138 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.874092 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.926396 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.955382 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 14:34:19 crc kubenswrapper[4817]: I1128 14:34:19.956071 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.078061 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.106351 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.316272 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.409146 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.426394 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.431842 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.487955 4817 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.585380 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.647347 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.674230 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.721777 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.855695 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.871553 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.911366 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 14:34:20 crc kubenswrapper[4817]: I1128 14:34:20.959046 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.005916 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.021016 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.032339 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.064813 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.074977 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.080142 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.120782 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.135561 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.165484 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.287696 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.306779 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.363565 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.415783 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.420190 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.452683 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.513197 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.555886 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.557159 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.658366 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.688548 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.716832 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.727237 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.737624 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.785210 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.818786 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.855123 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.887790 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.947626 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 14:34:21 crc kubenswrapper[4817]: I1128 14:34:21.998866 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.283786 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.321615 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.363504 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.384157 4817 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.404416 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.514283 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.611300 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.639001 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.682529 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.748250 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.817120 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.845667 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.971762 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.974090 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 14:34:22 crc kubenswrapper[4817]: I1128 14:34:22.989831 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.023344 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.250015 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.311280 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.319097 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.341037 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.431111 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.447075 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.487303 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.506366 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.652076 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.736789 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 14:34:23 crc kubenswrapper[4817]: I1128 14:34:23.855572 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.003659 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.048241 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.141855 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.156434 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.159522 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.165194 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.259657 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.413525 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.427626 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.475878 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.487434 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.487559 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.613393 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.621813 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.657455 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.723645 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.760048 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.940639 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 14:34:24 crc kubenswrapper[4817]: I1128 14:34:24.998464 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.022574 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.070562 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.144543 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.252979 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.262167 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.301789 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.336244 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.349933 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.389313 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.417042 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.468195 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.584179 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.764390 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.860614 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 14:34:25 crc kubenswrapper[4817]: I1128 14:34:25.915922 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.017731 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.269088 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.314317 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.332648 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.384949 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.407895 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.410170 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.450640 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.459797 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.485639 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.494368 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.499031 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.525175 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.534308 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.602875 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.629123 4817 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.653643 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.674285 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.684922 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.891178 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.908404 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.989527 4817 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.990541 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r22b9" podStartSLOduration=45.267206453 podStartE2EDuration="47.990513712s" podCreationTimestamp="2025-11-28 14:33:39 +0000 UTC" firstStartedPulling="2025-11-28 14:33:41.895204082 +0000 UTC m=+244.483182358" lastFinishedPulling="2025-11-28 14:33:44.618511351 +0000 UTC m=+247.206489617" observedRunningTime="2025-11-28 14:34:05.173984612 +0000 UTC m=+267.761962888" watchObservedRunningTime="2025-11-28 14:34:26.990513712 +0000 UTC m=+289.578492008" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.997656 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d7r7v" podStartSLOduration=46.549680384 podStartE2EDuration="48.997634896s" podCreationTimestamp="2025-11-28 14:33:38 +0000 UTC" firstStartedPulling="2025-11-28 14:33:39.868631408 +0000 UTC m=+242.456609674" lastFinishedPulling="2025-11-28 14:33:42.31658592 +0000 UTC m=+244.904564186" observedRunningTime="2025-11-28 14:34:05.306753866 +0000 UTC m=+267.894732152" watchObservedRunningTime="2025-11-28 14:34:26.997634896 +0000 UTC m=+289.585613202" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.998323 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-jczpw"] Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.998400 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-72n5z","openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 14:34:26 crc kubenswrapper[4817]: E1128 14:34:26.998752 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" containerName="oauth-openshift" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.998783 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" containerName="oauth-openshift" Nov 28 14:34:26 crc kubenswrapper[4817]: E1128 14:34:26.998815 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" containerName="installer" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.998828 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" containerName="installer" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.998886 4817 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.998924 4817 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="dd1a8a66-95ea-4dc3-b906-8fb49bb61503" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.999006 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" containerName="oauth-openshift" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.999028 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ce10d4-a7fa-491f-bbe2-50117f74f86c" containerName="installer" Nov 28 14:34:26 crc kubenswrapper[4817]: I1128 14:34:26.999700 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.007181 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.008497 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.008688 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.008840 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.008932 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.009303 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.009434 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.009509 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.009923 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.010158 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.010455 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.011597 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.012247 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.022330 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.029486 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.036855 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.049761 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.049820 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.049858 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.049936 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050003 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050037 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050077 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050116 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4s9r\" (UniqueName: \"kubernetes.io/projected/b466dce1-0ab0-4467-98d9-8f760ee011df-kube-api-access-p4s9r\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050259 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050366 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050404 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050428 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-dir\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050468 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-policies\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.050497 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.063030 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.063004242 podStartE2EDuration="21.063004242s" podCreationTimestamp="2025-11-28 14:34:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:34:27.055101437 +0000 UTC m=+289.643079753" watchObservedRunningTime="2025-11-28 14:34:27.063004242 +0000 UTC m=+289.650982518" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.135309 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.152142 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.153452 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.153643 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.153874 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.154025 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.154175 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.154333 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4s9r\" (UniqueName: \"kubernetes.io/projected/b466dce1-0ab0-4467-98d9-8f760ee011df-kube-api-access-p4s9r\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.154481 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.154638 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.154988 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.155135 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-dir\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.155285 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-policies\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.155422 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.155616 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.156598 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.157498 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-policies\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.157539 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.157573 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-dir\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.155341 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.159449 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.159899 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.161394 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.161830 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.163140 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.167648 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.170302 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.170973 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.173136 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4s9r\" (UniqueName: \"kubernetes.io/projected/b466dce1-0ab0-4467-98d9-8f760ee011df-kube-api-access-p4s9r\") pod \"oauth-openshift-64f9fb64bf-72n5z\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.179898 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.213350 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.305613 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.307973 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.309814 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.316749 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.332910 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.342840 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.447360 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.541284 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.571628 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.603013 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.653831 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.751599 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f890cb5e-2cff-4904-b9f5-3ecd702e869e" path="/var/lib/kubelet/pods/f890cb5e-2cff-4904-b9f5-3ecd702e869e/volumes" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.824280 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.826319 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.865228 4817 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.865537 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://e24185531a0806b3133ffed5328f254f3f4a2e17304e41d45c7d78d99178ed3f" gracePeriod=5 Nov 28 14:34:27 crc kubenswrapper[4817]: I1128 14:34:27.877645 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.063889 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.111691 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.145501 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.347196 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.458968 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.565900 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.594042 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.688853 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.814808 4817 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.864956 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 14:34:28 crc kubenswrapper[4817]: I1128 14:34:28.909265 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.046650 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.061276 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.126511 4817 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.443200 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.483534 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.502730 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.518449 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.638174 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.702435 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.907332 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 14:34:29 crc kubenswrapper[4817]: I1128 14:34:29.980958 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 14:34:30 crc kubenswrapper[4817]: I1128 14:34:30.059037 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 14:34:30 crc kubenswrapper[4817]: I1128 14:34:30.458889 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 14:34:30 crc kubenswrapper[4817]: I1128 14:34:30.615467 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 14:34:30 crc kubenswrapper[4817]: I1128 14:34:30.921015 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.253290 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.253751 4817 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="e24185531a0806b3133ffed5328f254f3f4a2e17304e41d45c7d78d99178ed3f" exitCode=137 Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.600756 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.600860 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656293 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656389 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656550 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656586 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656641 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656655 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656689 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656673 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.656836 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.657426 4817 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.657470 4817 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.657492 4817 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.657511 4817 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.671965 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.742305 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 28 14:34:33 crc kubenswrapper[4817]: I1128 14:34:33.758310 4817 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:34:34 crc kubenswrapper[4817]: I1128 14:34:34.267312 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 14:34:34 crc kubenswrapper[4817]: I1128 14:34:34.267421 4817 scope.go:117] "RemoveContainer" containerID="e24185531a0806b3133ffed5328f254f3f4a2e17304e41d45c7d78d99178ed3f" Nov 28 14:34:34 crc kubenswrapper[4817]: I1128 14:34:34.267511 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 14:34:40 crc kubenswrapper[4817]: I1128 14:34:40.236933 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 14:34:41 crc kubenswrapper[4817]: I1128 14:34:41.617299 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 14:34:42 crc kubenswrapper[4817]: I1128 14:34:42.040488 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 14:34:43 crc kubenswrapper[4817]: I1128 14:34:43.364444 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 14:34:43 crc kubenswrapper[4817]: I1128 14:34:43.891468 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 14:34:45 crc kubenswrapper[4817]: I1128 14:34:45.139050 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 14:34:45 crc kubenswrapper[4817]: I1128 14:34:45.900544 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 14:34:46 crc kubenswrapper[4817]: I1128 14:34:46.111277 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 14:34:48 crc kubenswrapper[4817]: I1128 14:34:48.342580 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 14:34:49 crc kubenswrapper[4817]: I1128 14:34:49.780122 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-72n5z"] Nov 28 14:34:50 crc kubenswrapper[4817]: I1128 14:34:50.280248 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-72n5z"] Nov 28 14:34:50 crc kubenswrapper[4817]: I1128 14:34:50.396423 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" event={"ID":"b466dce1-0ab0-4467-98d9-8f760ee011df","Type":"ContainerStarted","Data":"5433705f88eed1d2ba0b3aa62530087a86491603b974ece813b0d211a4502ffd"} Nov 28 14:34:51 crc kubenswrapper[4817]: I1128 14:34:51.138218 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 14:34:51 crc kubenswrapper[4817]: I1128 14:34:51.405465 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" event={"ID":"b466dce1-0ab0-4467-98d9-8f760ee011df","Type":"ContainerStarted","Data":"a4222f595ad08222eabce90b8a322daf30742aa73200349eae2ddc50719934d6"} Nov 28 14:34:51 crc kubenswrapper[4817]: I1128 14:34:51.405780 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:51 crc kubenswrapper[4817]: I1128 14:34:51.415602 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:34:51 crc kubenswrapper[4817]: I1128 14:34:51.447299 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" podStartSLOduration=78.447233614 podStartE2EDuration="1m18.447233614s" podCreationTimestamp="2025-11-28 14:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:34:51.44012675 +0000 UTC m=+314.028105066" watchObservedRunningTime="2025-11-28 14:34:51.447233614 +0000 UTC m=+314.035211920" Nov 28 14:34:52 crc kubenswrapper[4817]: I1128 14:34:52.911950 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 14:34:54 crc kubenswrapper[4817]: I1128 14:34:54.045756 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 14:34:54 crc kubenswrapper[4817]: I1128 14:34:54.778706 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 14:34:56 crc kubenswrapper[4817]: I1128 14:34:56.707084 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 14:34:56 crc kubenswrapper[4817]: I1128 14:34:56.867759 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 14:34:58 crc kubenswrapper[4817]: I1128 14:34:58.369845 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 14:34:58 crc kubenswrapper[4817]: I1128 14:34:58.861688 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 14:34:59 crc kubenswrapper[4817]: I1128 14:34:59.419714 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 14:35:00 crc kubenswrapper[4817]: I1128 14:35:00.753157 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 14:35:01 crc kubenswrapper[4817]: I1128 14:35:01.759196 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 14:35:06 crc kubenswrapper[4817]: I1128 14:35:06.610040 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.291929 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kzstg"] Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.292708 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" podUID="a4552569-578c-4694-8e3b-1a3ec68380e9" containerName="controller-manager" containerID="cri-o://1d068147cbbf0344fc9f546245c4f2fbc1bcc6db806a51b5fa0f7721ca0d40a9" gracePeriod=30 Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.394210 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d"] Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.394445 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" podUID="67eed89a-6d2c-4b98-bf5b-13431b920d0c" containerName="route-controller-manager" containerID="cri-o://3910f73f0a2b063c292f32612e970b794daaef7400184c7652da212a453d687b" gracePeriod=30 Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.610372 4817 generic.go:334] "Generic (PLEG): container finished" podID="67eed89a-6d2c-4b98-bf5b-13431b920d0c" containerID="3910f73f0a2b063c292f32612e970b794daaef7400184c7652da212a453d687b" exitCode=0 Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.610476 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" event={"ID":"67eed89a-6d2c-4b98-bf5b-13431b920d0c","Type":"ContainerDied","Data":"3910f73f0a2b063c292f32612e970b794daaef7400184c7652da212a453d687b"} Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.613050 4817 generic.go:334] "Generic (PLEG): container finished" podID="a4552569-578c-4694-8e3b-1a3ec68380e9" containerID="1d068147cbbf0344fc9f546245c4f2fbc1bcc6db806a51b5fa0f7721ca0d40a9" exitCode=0 Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.613078 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" event={"ID":"a4552569-578c-4694-8e3b-1a3ec68380e9","Type":"ContainerDied","Data":"1d068147cbbf0344fc9f546245c4f2fbc1bcc6db806a51b5fa0f7721ca0d40a9"} Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.613094 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" event={"ID":"a4552569-578c-4694-8e3b-1a3ec68380e9","Type":"ContainerDied","Data":"7579eba1e499e0ab1681fe27ce12b818990e3be54f5f38dcf170230482a22719"} Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.613104 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7579eba1e499e0ab1681fe27ce12b818990e3be54f5f38dcf170230482a22719" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.630221 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.650163 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4552569-578c-4694-8e3b-1a3ec68380e9-serving-cert\") pod \"a4552569-578c-4694-8e3b-1a3ec68380e9\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.650265 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-config\") pod \"a4552569-578c-4694-8e3b-1a3ec68380e9\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.650291 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-proxy-ca-bundles\") pod \"a4552569-578c-4694-8e3b-1a3ec68380e9\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.650315 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trq9b\" (UniqueName: \"kubernetes.io/projected/a4552569-578c-4694-8e3b-1a3ec68380e9-kube-api-access-trq9b\") pod \"a4552569-578c-4694-8e3b-1a3ec68380e9\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.650402 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-client-ca\") pod \"a4552569-578c-4694-8e3b-1a3ec68380e9\" (UID: \"a4552569-578c-4694-8e3b-1a3ec68380e9\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.651692 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-client-ca" (OuterVolumeSpecName: "client-ca") pod "a4552569-578c-4694-8e3b-1a3ec68380e9" (UID: "a4552569-578c-4694-8e3b-1a3ec68380e9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.651805 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-config" (OuterVolumeSpecName: "config") pod "a4552569-578c-4694-8e3b-1a3ec68380e9" (UID: "a4552569-578c-4694-8e3b-1a3ec68380e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.652375 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a4552569-578c-4694-8e3b-1a3ec68380e9" (UID: "a4552569-578c-4694-8e3b-1a3ec68380e9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.659054 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4552569-578c-4694-8e3b-1a3ec68380e9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a4552569-578c-4694-8e3b-1a3ec68380e9" (UID: "a4552569-578c-4694-8e3b-1a3ec68380e9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.659081 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4552569-578c-4694-8e3b-1a3ec68380e9-kube-api-access-trq9b" (OuterVolumeSpecName: "kube-api-access-trq9b") pod "a4552569-578c-4694-8e3b-1a3ec68380e9" (UID: "a4552569-578c-4694-8e3b-1a3ec68380e9"). InnerVolumeSpecName "kube-api-access-trq9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.684680 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.750914 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eed89a-6d2c-4b98-bf5b-13431b920d0c-serving-cert\") pod \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.750975 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-client-ca\") pod \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751045 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-config\") pod \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751100 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwcfm\" (UniqueName: \"kubernetes.io/projected/67eed89a-6d2c-4b98-bf5b-13431b920d0c-kube-api-access-dwcfm\") pod \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\" (UID: \"67eed89a-6d2c-4b98-bf5b-13431b920d0c\") " Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751335 4817 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751349 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4552569-578c-4694-8e3b-1a3ec68380e9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751358 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751478 4817 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4552569-578c-4694-8e3b-1a3ec68380e9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751652 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trq9b\" (UniqueName: \"kubernetes.io/projected/a4552569-578c-4694-8e3b-1a3ec68380e9-kube-api-access-trq9b\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751836 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-config" (OuterVolumeSpecName: "config") pod "67eed89a-6d2c-4b98-bf5b-13431b920d0c" (UID: "67eed89a-6d2c-4b98-bf5b-13431b920d0c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.751872 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-client-ca" (OuterVolumeSpecName: "client-ca") pod "67eed89a-6d2c-4b98-bf5b-13431b920d0c" (UID: "67eed89a-6d2c-4b98-bf5b-13431b920d0c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.754312 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67eed89a-6d2c-4b98-bf5b-13431b920d0c-kube-api-access-dwcfm" (OuterVolumeSpecName: "kube-api-access-dwcfm") pod "67eed89a-6d2c-4b98-bf5b-13431b920d0c" (UID: "67eed89a-6d2c-4b98-bf5b-13431b920d0c"). InnerVolumeSpecName "kube-api-access-dwcfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.768089 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67eed89a-6d2c-4b98-bf5b-13431b920d0c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "67eed89a-6d2c-4b98-bf5b-13431b920d0c" (UID: "67eed89a-6d2c-4b98-bf5b-13431b920d0c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.852962 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.853269 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwcfm\" (UniqueName: \"kubernetes.io/projected/67eed89a-6d2c-4b98-bf5b-13431b920d0c-kube-api-access-dwcfm\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.853285 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eed89a-6d2c-4b98-bf5b-13431b920d0c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:23 crc kubenswrapper[4817]: I1128 14:35:23.853299 4817 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67eed89a-6d2c-4b98-bf5b-13431b920d0c-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.620466 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kzstg" Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.620472 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" event={"ID":"67eed89a-6d2c-4b98-bf5b-13431b920d0c","Type":"ContainerDied","Data":"83c4fce7ec7a9d45ead6a24fde7e69e67855735043befed228189f978d1fb64b"} Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.620546 4817 scope.go:117] "RemoveContainer" containerID="3910f73f0a2b063c292f32612e970b794daaef7400184c7652da212a453d687b" Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.620466 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d" Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.641395 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kzstg"] Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.649279 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kzstg"] Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.655309 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d"] Nov 28 14:35:24 crc kubenswrapper[4817]: I1128 14:35:24.660546 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tfg8d"] Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.239458 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b999966dd-7j74p"] Nov 28 14:35:25 crc kubenswrapper[4817]: E1128 14:35:25.239831 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4552569-578c-4694-8e3b-1a3ec68380e9" containerName="controller-manager" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.239859 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4552569-578c-4694-8e3b-1a3ec68380e9" containerName="controller-manager" Nov 28 14:35:25 crc kubenswrapper[4817]: E1128 14:35:25.239881 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.239892 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 14:35:25 crc kubenswrapper[4817]: E1128 14:35:25.239908 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67eed89a-6d2c-4b98-bf5b-13431b920d0c" containerName="route-controller-manager" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.239921 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="67eed89a-6d2c-4b98-bf5b-13431b920d0c" containerName="route-controller-manager" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.240080 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4552569-578c-4694-8e3b-1a3ec68380e9" containerName="controller-manager" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.240108 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.240127 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="67eed89a-6d2c-4b98-bf5b-13431b920d0c" containerName="route-controller-manager" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.240685 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.246945 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.247687 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr"] Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.248892 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.248902 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.249051 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.249648 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.250017 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.250052 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.253944 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b999966dd-7j74p"] Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.257430 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr"] Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.258293 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.258325 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.258486 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.258858 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.259003 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.260448 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.261625 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.268969 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-proxy-ca-bundles\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.271897 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-serving-cert\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.271994 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jkzr\" (UniqueName: \"kubernetes.io/projected/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-kube-api-access-9jkzr\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.272060 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-config\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.272117 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-client-ca\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.272149 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-client-ca\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.272170 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-config\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.272204 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-serving-cert\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.272246 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5svvs\" (UniqueName: \"kubernetes.io/projected/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-kube-api-access-5svvs\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373268 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-serving-cert\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373323 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jkzr\" (UniqueName: \"kubernetes.io/projected/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-kube-api-access-9jkzr\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373352 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-config\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373370 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-client-ca\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373388 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-client-ca\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373404 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-config\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373426 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-serving-cert\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373452 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5svvs\" (UniqueName: \"kubernetes.io/projected/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-kube-api-access-5svvs\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.373479 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-proxy-ca-bundles\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.374511 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-client-ca\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.374543 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-proxy-ca-bundles\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.375133 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-client-ca\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.375519 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-config\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.375920 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-config\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.378915 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-serving-cert\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.379399 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-serving-cert\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.395464 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jkzr\" (UniqueName: \"kubernetes.io/projected/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-kube-api-access-9jkzr\") pod \"route-controller-manager-588f69b4f6-bhlhr\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.405383 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5svvs\" (UniqueName: \"kubernetes.io/projected/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-kube-api-access-5svvs\") pod \"controller-manager-b999966dd-7j74p\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.560954 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.569253 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.746061 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67eed89a-6d2c-4b98-bf5b-13431b920d0c" path="/var/lib/kubelet/pods/67eed89a-6d2c-4b98-bf5b-13431b920d0c/volumes" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.747051 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4552569-578c-4694-8e3b-1a3ec68380e9" path="/var/lib/kubelet/pods/a4552569-578c-4694-8e3b-1a3ec68380e9/volumes" Nov 28 14:35:25 crc kubenswrapper[4817]: I1128 14:35:25.804914 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b999966dd-7j74p"] Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.040792 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr"] Nov 28 14:35:26 crc kubenswrapper[4817]: W1128 14:35:26.043205 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1adde42f_3cbb_49e5_b8ba_e35fbd42db21.slice/crio-fb27ef5497bceb09c7fb16e2956a3465b999d87f31774c8978c6a1d59022439c WatchSource:0}: Error finding container fb27ef5497bceb09c7fb16e2956a3465b999d87f31774c8978c6a1d59022439c: Status 404 returned error can't find the container with id fb27ef5497bceb09c7fb16e2956a3465b999d87f31774c8978c6a1d59022439c Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.633263 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" event={"ID":"1adde42f-3cbb-49e5-b8ba-e35fbd42db21","Type":"ContainerStarted","Data":"ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21"} Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.633309 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" event={"ID":"1adde42f-3cbb-49e5-b8ba-e35fbd42db21","Type":"ContainerStarted","Data":"fb27ef5497bceb09c7fb16e2956a3465b999d87f31774c8978c6a1d59022439c"} Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.633534 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.635548 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" event={"ID":"8fad6988-4bfb-48c4-9775-82cf2bbc5d75","Type":"ContainerStarted","Data":"e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f"} Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.635599 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" event={"ID":"8fad6988-4bfb-48c4-9775-82cf2bbc5d75","Type":"ContainerStarted","Data":"d3ea9bb5c60f017713c557ba509ff08041ceb7f1e53c55a430b2bccf4cdbd041"} Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.635972 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.640605 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.657879 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" podStartSLOduration=3.657855962 podStartE2EDuration="3.657855962s" podCreationTimestamp="2025-11-28 14:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:35:26.651296913 +0000 UTC m=+349.239275169" watchObservedRunningTime="2025-11-28 14:35:26.657855962 +0000 UTC m=+349.245834258" Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.668002 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" podStartSLOduration=3.6679819719999998 podStartE2EDuration="3.667981972s" podCreationTimestamp="2025-11-28 14:35:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:35:26.666956255 +0000 UTC m=+349.254934521" watchObservedRunningTime="2025-11-28 14:35:26.667981972 +0000 UTC m=+349.255960238" Nov 28 14:35:26 crc kubenswrapper[4817]: I1128 14:35:26.825114 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:30 crc kubenswrapper[4817]: I1128 14:35:30.611519 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b999966dd-7j74p"] Nov 28 14:35:30 crc kubenswrapper[4817]: I1128 14:35:30.612204 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" podUID="8fad6988-4bfb-48c4-9775-82cf2bbc5d75" containerName="controller-manager" containerID="cri-o://e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f" gracePeriod=30 Nov 28 14:35:30 crc kubenswrapper[4817]: I1128 14:35:30.647769 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr"] Nov 28 14:35:30 crc kubenswrapper[4817]: I1128 14:35:30.647968 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" podUID="1adde42f-3cbb-49e5-b8ba-e35fbd42db21" containerName="route-controller-manager" containerID="cri-o://ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21" gracePeriod=30 Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.141202 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.146813 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274360 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-serving-cert\") pod \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274440 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5svvs\" (UniqueName: \"kubernetes.io/projected/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-kube-api-access-5svvs\") pod \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274491 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-proxy-ca-bundles\") pod \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274515 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jkzr\" (UniqueName: \"kubernetes.io/projected/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-kube-api-access-9jkzr\") pod \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274531 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-client-ca\") pod \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274608 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-config\") pod \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274628 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-config\") pod \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274643 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-serving-cert\") pod \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\" (UID: \"8fad6988-4bfb-48c4-9775-82cf2bbc5d75\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.274672 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-client-ca\") pod \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\" (UID: \"1adde42f-3cbb-49e5-b8ba-e35fbd42db21\") " Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.275269 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-client-ca" (OuterVolumeSpecName: "client-ca") pod "8fad6988-4bfb-48c4-9775-82cf2bbc5d75" (UID: "8fad6988-4bfb-48c4-9775-82cf2bbc5d75"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.275287 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-client-ca" (OuterVolumeSpecName: "client-ca") pod "1adde42f-3cbb-49e5-b8ba-e35fbd42db21" (UID: "1adde42f-3cbb-49e5-b8ba-e35fbd42db21"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.275320 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8fad6988-4bfb-48c4-9775-82cf2bbc5d75" (UID: "8fad6988-4bfb-48c4-9775-82cf2bbc5d75"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.275322 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-config" (OuterVolumeSpecName: "config") pod "1adde42f-3cbb-49e5-b8ba-e35fbd42db21" (UID: "1adde42f-3cbb-49e5-b8ba-e35fbd42db21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.275374 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-config" (OuterVolumeSpecName: "config") pod "8fad6988-4bfb-48c4-9775-82cf2bbc5d75" (UID: "8fad6988-4bfb-48c4-9775-82cf2bbc5d75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.280048 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-kube-api-access-5svvs" (OuterVolumeSpecName: "kube-api-access-5svvs") pod "8fad6988-4bfb-48c4-9775-82cf2bbc5d75" (UID: "8fad6988-4bfb-48c4-9775-82cf2bbc5d75"). InnerVolumeSpecName "kube-api-access-5svvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.280122 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8fad6988-4bfb-48c4-9775-82cf2bbc5d75" (UID: "8fad6988-4bfb-48c4-9775-82cf2bbc5d75"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.281289 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-kube-api-access-9jkzr" (OuterVolumeSpecName: "kube-api-access-9jkzr") pod "1adde42f-3cbb-49e5-b8ba-e35fbd42db21" (UID: "1adde42f-3cbb-49e5-b8ba-e35fbd42db21"). InnerVolumeSpecName "kube-api-access-9jkzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.286386 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1adde42f-3cbb-49e5-b8ba-e35fbd42db21" (UID: "1adde42f-3cbb-49e5-b8ba-e35fbd42db21"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.377810 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378146 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378269 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378389 4817 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378495 4817 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378601 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5svvs\" (UniqueName: \"kubernetes.io/projected/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-kube-api-access-5svvs\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378707 4817 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.378872 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jkzr\" (UniqueName: \"kubernetes.io/projected/1adde42f-3cbb-49e5-b8ba-e35fbd42db21-kube-api-access-9jkzr\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.379049 4817 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fad6988-4bfb-48c4-9775-82cf2bbc5d75-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.668225 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fad6988-4bfb-48c4-9775-82cf2bbc5d75" containerID="e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f" exitCode=0 Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.668290 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" event={"ID":"8fad6988-4bfb-48c4-9775-82cf2bbc5d75","Type":"ContainerDied","Data":"e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f"} Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.668324 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.668379 4817 scope.go:117] "RemoveContainer" containerID="e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.668354 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b999966dd-7j74p" event={"ID":"8fad6988-4bfb-48c4-9775-82cf2bbc5d75","Type":"ContainerDied","Data":"d3ea9bb5c60f017713c557ba509ff08041ceb7f1e53c55a430b2bccf4cdbd041"} Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.672919 4817 generic.go:334] "Generic (PLEG): container finished" podID="1adde42f-3cbb-49e5-b8ba-e35fbd42db21" containerID="ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21" exitCode=0 Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.673028 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" event={"ID":"1adde42f-3cbb-49e5-b8ba-e35fbd42db21","Type":"ContainerDied","Data":"ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21"} Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.673074 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" event={"ID":"1adde42f-3cbb-49e5-b8ba-e35fbd42db21","Type":"ContainerDied","Data":"fb27ef5497bceb09c7fb16e2956a3465b999d87f31774c8978c6a1d59022439c"} Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.673171 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.718927 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b999966dd-7j74p"] Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.723524 4817 scope.go:117] "RemoveContainer" containerID="e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f" Nov 28 14:35:31 crc kubenswrapper[4817]: E1128 14:35:31.724361 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f\": container with ID starting with e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f not found: ID does not exist" containerID="e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.724415 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f"} err="failed to get container status \"e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f\": rpc error: code = NotFound desc = could not find container \"e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f\": container with ID starting with e7e4e279c4a4aa97970eb3d57f8c542bff0596b2515798a2a7aafd8a3f86812f not found: ID does not exist" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.724442 4817 scope.go:117] "RemoveContainer" containerID="ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.755043 4817 scope.go:117] "RemoveContainer" containerID="ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21" Nov 28 14:35:31 crc kubenswrapper[4817]: E1128 14:35:31.756124 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21\": container with ID starting with ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21 not found: ID does not exist" containerID="ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.756191 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21"} err="failed to get container status \"ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21\": rpc error: code = NotFound desc = could not find container \"ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21\": container with ID starting with ff7efd15e21f382bbbd2eac4787e4044ea9608d126cdd810fd7a5af88e110d21 not found: ID does not exist" Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.770472 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b999966dd-7j74p"] Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.770646 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr"] Nov 28 14:35:31 crc kubenswrapper[4817]: I1128 14:35:31.770681 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588f69b4f6-bhlhr"] Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.243297 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk"] Nov 28 14:35:32 crc kubenswrapper[4817]: E1128 14:35:32.243873 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fad6988-4bfb-48c4-9775-82cf2bbc5d75" containerName="controller-manager" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.243892 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fad6988-4bfb-48c4-9775-82cf2bbc5d75" containerName="controller-manager" Nov 28 14:35:32 crc kubenswrapper[4817]: E1128 14:35:32.243912 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1adde42f-3cbb-49e5-b8ba-e35fbd42db21" containerName="route-controller-manager" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.243921 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="1adde42f-3cbb-49e5-b8ba-e35fbd42db21" containerName="route-controller-manager" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.244013 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fad6988-4bfb-48c4-9775-82cf2bbc5d75" containerName="controller-manager" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.244031 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="1adde42f-3cbb-49e5-b8ba-e35fbd42db21" containerName="route-controller-manager" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.244390 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.246497 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.246631 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6dd965b4d7-9457v"] Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.246688 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.246928 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.247023 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.247027 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.247036 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.247326 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.250877 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.251031 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.251181 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.251260 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.251531 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.251844 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.259848 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.263063 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6dd965b4d7-9457v"] Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.269514 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk"] Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289372 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-serving-cert\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289407 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-config\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289425 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69ltw\" (UniqueName: \"kubernetes.io/projected/ee7f3c5f-836f-4370-8b47-1e1046314988-kube-api-access-69ltw\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289452 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-config\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289570 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-client-ca\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289618 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jkts\" (UniqueName: \"kubernetes.io/projected/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-kube-api-access-2jkts\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289656 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-proxy-ca-bundles\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289694 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-client-ca\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.289843 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee7f3c5f-836f-4370-8b47-1e1046314988-serving-cert\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391071 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-client-ca\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391134 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee7f3c5f-836f-4370-8b47-1e1046314988-serving-cert\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391219 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-serving-cert\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391245 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-config\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391269 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69ltw\" (UniqueName: \"kubernetes.io/projected/ee7f3c5f-836f-4370-8b47-1e1046314988-kube-api-access-69ltw\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391305 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-config\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391329 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-client-ca\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391355 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jkts\" (UniqueName: \"kubernetes.io/projected/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-kube-api-access-2jkts\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391380 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-proxy-ca-bundles\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.391954 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-client-ca\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.392304 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-proxy-ca-bundles\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.393606 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-config\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.394126 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-client-ca\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.394648 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7f3c5f-836f-4370-8b47-1e1046314988-config\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.400391 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-serving-cert\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.400912 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee7f3c5f-836f-4370-8b47-1e1046314988-serving-cert\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.408958 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69ltw\" (UniqueName: \"kubernetes.io/projected/ee7f3c5f-836f-4370-8b47-1e1046314988-kube-api-access-69ltw\") pod \"controller-manager-6dd965b4d7-9457v\" (UID: \"ee7f3c5f-836f-4370-8b47-1e1046314988\") " pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.411320 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jkts\" (UniqueName: \"kubernetes.io/projected/8a5f03bb-3d6c-4521-987d-f4e72551d4fb-kube-api-access-2jkts\") pod \"route-controller-manager-76b94c565-rptmk\" (UID: \"8a5f03bb-3d6c-4521-987d-f4e72551d4fb\") " pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.575644 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.583623 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.768445 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk"] Nov 28 14:35:32 crc kubenswrapper[4817]: I1128 14:35:32.872848 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6dd965b4d7-9457v"] Nov 28 14:35:32 crc kubenswrapper[4817]: W1128 14:35:32.904574 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee7f3c5f_836f_4370_8b47_1e1046314988.slice/crio-b979608e5651c007766b4b2be7aa57527a72d567e3c293ac66de778b4b873e3e WatchSource:0}: Error finding container b979608e5651c007766b4b2be7aa57527a72d567e3c293ac66de778b4b873e3e: Status 404 returned error can't find the container with id b979608e5651c007766b4b2be7aa57527a72d567e3c293ac66de778b4b873e3e Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.694355 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" event={"ID":"ee7f3c5f-836f-4370-8b47-1e1046314988","Type":"ContainerStarted","Data":"66d5956ea0e5d659ee77fe7d992e1e432da132a3bf2233a8dae339b4030e70eb"} Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.694755 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.694779 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" event={"ID":"ee7f3c5f-836f-4370-8b47-1e1046314988","Type":"ContainerStarted","Data":"b979608e5651c007766b4b2be7aa57527a72d567e3c293ac66de778b4b873e3e"} Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.696339 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" event={"ID":"8a5f03bb-3d6c-4521-987d-f4e72551d4fb","Type":"ContainerStarted","Data":"a53593c2bfaa8962b558566766083ae16e61808e202cd71cf2800fba2bb37af3"} Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.696363 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" event={"ID":"8a5f03bb-3d6c-4521-987d-f4e72551d4fb","Type":"ContainerStarted","Data":"8be36c31fbd06cea31c527b83469a6149f51fe8cd3c607aa06c80504087bb59b"} Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.696588 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.698565 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.700760 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.710488 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6dd965b4d7-9457v" podStartSLOduration=3.710474465 podStartE2EDuration="3.710474465s" podCreationTimestamp="2025-11-28 14:35:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:35:33.710057364 +0000 UTC m=+356.298035630" watchObservedRunningTime="2025-11-28 14:35:33.710474465 +0000 UTC m=+356.298452731" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.737376 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-76b94c565-rptmk" podStartSLOduration=3.737356265 podStartE2EDuration="3.737356265s" podCreationTimestamp="2025-11-28 14:35:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:35:33.733445105 +0000 UTC m=+356.321423371" watchObservedRunningTime="2025-11-28 14:35:33.737356265 +0000 UTC m=+356.325334531" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.746077 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1adde42f-3cbb-49e5-b8ba-e35fbd42db21" path="/var/lib/kubelet/pods/1adde42f-3cbb-49e5-b8ba-e35fbd42db21/volumes" Nov 28 14:35:33 crc kubenswrapper[4817]: I1128 14:35:33.746833 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fad6988-4bfb-48c4-9775-82cf2bbc5d75" path="/var/lib/kubelet/pods/8fad6988-4bfb-48c4-9775-82cf2bbc5d75/volumes" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.274666 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6846j"] Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.276077 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.289548 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6846j"] Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324474 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b076396-42a7-4330-acf1-ecae9f28b438-registry-certificates\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324519 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b076396-42a7-4330-acf1-ecae9f28b438-trusted-ca\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324538 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b076396-42a7-4330-acf1-ecae9f28b438-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324581 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324612 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-bound-sa-token\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324640 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2fv\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-kube-api-access-nx2fv\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324658 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b076396-42a7-4330-acf1-ecae9f28b438-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.324679 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-registry-tls\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.388277 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.425959 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-bound-sa-token\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426022 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx2fv\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-kube-api-access-nx2fv\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426051 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b076396-42a7-4330-acf1-ecae9f28b438-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426074 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-registry-tls\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426101 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b076396-42a7-4330-acf1-ecae9f28b438-registry-certificates\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426118 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b076396-42a7-4330-acf1-ecae9f28b438-trusted-ca\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426136 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b076396-42a7-4330-acf1-ecae9f28b438-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.426592 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b076396-42a7-4330-acf1-ecae9f28b438-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.427331 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b076396-42a7-4330-acf1-ecae9f28b438-registry-certificates\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.428143 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b076396-42a7-4330-acf1-ecae9f28b438-trusted-ca\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.432545 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-registry-tls\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.436139 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b076396-42a7-4330-acf1-ecae9f28b438-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.442086 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-bound-sa-token\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.449046 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx2fv\" (UniqueName: \"kubernetes.io/projected/1b076396-42a7-4330-acf1-ecae9f28b438-kube-api-access-nx2fv\") pod \"image-registry-66df7c8f76-6846j\" (UID: \"1b076396-42a7-4330-acf1-ecae9f28b438\") " pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:49 crc kubenswrapper[4817]: I1128 14:35:49.593244 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:50 crc kubenswrapper[4817]: I1128 14:35:50.072854 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6846j"] Nov 28 14:35:50 crc kubenswrapper[4817]: W1128 14:35:50.089853 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b076396_42a7_4330_acf1_ecae9f28b438.slice/crio-85b4fa9df1bf79ce5ebc8b137eb7a6abcbb6deca30e3ceada6356bbac9420d16 WatchSource:0}: Error finding container 85b4fa9df1bf79ce5ebc8b137eb7a6abcbb6deca30e3ceada6356bbac9420d16: Status 404 returned error can't find the container with id 85b4fa9df1bf79ce5ebc8b137eb7a6abcbb6deca30e3ceada6356bbac9420d16 Nov 28 14:35:50 crc kubenswrapper[4817]: I1128 14:35:50.815998 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" event={"ID":"1b076396-42a7-4330-acf1-ecae9f28b438","Type":"ContainerStarted","Data":"d1019a5eae0f4a74f8a3c748de25493eb730c0ec792d88f23da558042c0841be"} Nov 28 14:35:50 crc kubenswrapper[4817]: I1128 14:35:50.816047 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" event={"ID":"1b076396-42a7-4330-acf1-ecae9f28b438","Type":"ContainerStarted","Data":"85b4fa9df1bf79ce5ebc8b137eb7a6abcbb6deca30e3ceada6356bbac9420d16"} Nov 28 14:35:50 crc kubenswrapper[4817]: I1128 14:35:50.845054 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" podStartSLOduration=1.84503599 podStartE2EDuration="1.84503599s" podCreationTimestamp="2025-11-28 14:35:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:35:50.843414618 +0000 UTC m=+373.431392904" watchObservedRunningTime="2025-11-28 14:35:50.84503599 +0000 UTC m=+373.433014256" Nov 28 14:35:51 crc kubenswrapper[4817]: I1128 14:35:51.567894 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-72n5z"] Nov 28 14:35:51 crc kubenswrapper[4817]: I1128 14:35:51.821242 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:35:54 crc kubenswrapper[4817]: I1128 14:35:54.045338 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:35:54 crc kubenswrapper[4817]: I1128 14:35:54.045428 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:36:09 crc kubenswrapper[4817]: I1128 14:36:09.602596 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6846j" Nov 28 14:36:09 crc kubenswrapper[4817]: I1128 14:36:09.674935 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v76hp"] Nov 28 14:36:16 crc kubenswrapper[4817]: I1128 14:36:16.598020 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" podUID="b466dce1-0ab0-4467-98d9-8f760ee011df" containerName="oauth-openshift" containerID="cri-o://a4222f595ad08222eabce90b8a322daf30742aa73200349eae2ddc50719934d6" gracePeriod=15 Nov 28 14:36:16 crc kubenswrapper[4817]: I1128 14:36:16.983318 4817 generic.go:334] "Generic (PLEG): container finished" podID="b466dce1-0ab0-4467-98d9-8f760ee011df" containerID="a4222f595ad08222eabce90b8a322daf30742aa73200349eae2ddc50719934d6" exitCode=0 Nov 28 14:36:16 crc kubenswrapper[4817]: I1128 14:36:16.983413 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" event={"ID":"b466dce1-0ab0-4467-98d9-8f760ee011df","Type":"ContainerDied","Data":"a4222f595ad08222eabce90b8a322daf30742aa73200349eae2ddc50719934d6"} Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.017802 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.067916 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7854b76c6-rhmh8"] Nov 28 14:36:17 crc kubenswrapper[4817]: E1128 14:36:17.068342 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b466dce1-0ab0-4467-98d9-8f760ee011df" containerName="oauth-openshift" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.068384 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b466dce1-0ab0-4467-98d9-8f760ee011df" containerName="oauth-openshift" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.068602 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b466dce1-0ab0-4467-98d9-8f760ee011df" containerName="oauth-openshift" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.070677 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.072850 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7854b76c6-rhmh8"] Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148411 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-service-ca\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148487 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-router-certs\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148524 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-serving-cert\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148563 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-trusted-ca-bundle\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148596 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-cliconfig\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148632 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-policies\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148657 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-ocp-branding-template\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148691 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4s9r\" (UniqueName: \"kubernetes.io/projected/b466dce1-0ab0-4467-98d9-8f760ee011df-kube-api-access-p4s9r\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148724 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-provider-selection\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148811 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-session\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148848 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-idp-0-file-data\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148875 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-login\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148901 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-error\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.148936 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-dir\") pod \"b466dce1-0ab0-4467-98d9-8f760ee011df\" (UID: \"b466dce1-0ab0-4467-98d9-8f760ee011df\") " Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149142 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149186 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-audit-policies\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149218 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-login\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149241 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149266 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-error\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149287 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149318 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149343 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/088be55a-5151-48c8-baad-8e0e00eebcc7-audit-dir\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149378 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-session\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149420 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149453 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldkwh\" (UniqueName: \"kubernetes.io/projected/088be55a-5151-48c8-baad-8e0e00eebcc7-kube-api-access-ldkwh\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149488 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149521 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149548 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149630 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.149660 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.150006 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.150802 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.151004 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.154513 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.154607 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.154638 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.154777 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.154936 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.156053 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.156424 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b466dce1-0ab0-4467-98d9-8f760ee011df-kube-api-access-p4s9r" (OuterVolumeSpecName: "kube-api-access-p4s9r") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "kube-api-access-p4s9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.156629 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.159602 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b466dce1-0ab0-4467-98d9-8f760ee011df" (UID: "b466dce1-0ab0-4467-98d9-8f760ee011df"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.250998 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251110 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-audit-policies\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251155 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-login\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251185 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251222 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-error\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251253 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251294 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251336 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/088be55a-5151-48c8-baad-8e0e00eebcc7-audit-dir\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251375 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-session\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251421 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251452 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldkwh\" (UniqueName: \"kubernetes.io/projected/088be55a-5151-48c8-baad-8e0e00eebcc7-kube-api-access-ldkwh\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251494 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251535 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251575 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251667 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251688 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4s9r\" (UniqueName: \"kubernetes.io/projected/b466dce1-0ab0-4467-98d9-8f760ee011df-kube-api-access-p4s9r\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251710 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251763 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251783 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251802 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251821 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251842 4817 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251865 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251884 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251902 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251921 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251944 4817 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.251962 4817 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b466dce1-0ab0-4467-98d9-8f760ee011df-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.252248 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/088be55a-5151-48c8-baad-8e0e00eebcc7-audit-dir\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.252564 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.252883 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.254218 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.254434 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/088be55a-5151-48c8-baad-8e0e00eebcc7-audit-policies\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.255352 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-error\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.256029 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.256992 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-login\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.257712 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.258488 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.260125 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-session\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.260933 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.261869 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/088be55a-5151-48c8-baad-8e0e00eebcc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.277112 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldkwh\" (UniqueName: \"kubernetes.io/projected/088be55a-5151-48c8-baad-8e0e00eebcc7-kube-api-access-ldkwh\") pod \"oauth-openshift-7854b76c6-rhmh8\" (UID: \"088be55a-5151-48c8-baad-8e0e00eebcc7\") " pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.402698 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.761989 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7854b76c6-rhmh8"] Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.991452 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" event={"ID":"b466dce1-0ab0-4467-98d9-8f760ee011df","Type":"ContainerDied","Data":"5433705f88eed1d2ba0b3aa62530087a86491603b974ece813b0d211a4502ffd"} Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.991461 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-72n5z" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.991507 4817 scope.go:117] "RemoveContainer" containerID="a4222f595ad08222eabce90b8a322daf30742aa73200349eae2ddc50719934d6" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.995090 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" event={"ID":"088be55a-5151-48c8-baad-8e0e00eebcc7","Type":"ContainerStarted","Data":"f6587ff7b3025239268c4723f70c41429561c67779ef6f1c7ca8bd5e979486af"} Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.995137 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" event={"ID":"088be55a-5151-48c8-baad-8e0e00eebcc7","Type":"ContainerStarted","Data":"6428ce2d8660f7136399d00f5e23dac1f6c1ad6e7dfdae3054e1c3d5343f2bfc"} Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.995644 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.998010 4817 patch_prober.go:28] interesting pod/oauth-openshift-7854b76c6-rhmh8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" start-of-body= Nov 28 14:36:17 crc kubenswrapper[4817]: I1128 14:36:17.998047 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" podUID="088be55a-5151-48c8-baad-8e0e00eebcc7" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" Nov 28 14:36:18 crc kubenswrapper[4817]: I1128 14:36:18.022142 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-72n5z"] Nov 28 14:36:18 crc kubenswrapper[4817]: I1128 14:36:18.027284 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-72n5z"] Nov 28 14:36:18 crc kubenswrapper[4817]: I1128 14:36:18.033416 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" podStartSLOduration=27.033403427 podStartE2EDuration="27.033403427s" podCreationTimestamp="2025-11-28 14:35:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:36:18.02962672 +0000 UTC m=+400.617604986" watchObservedRunningTime="2025-11-28 14:36:18.033403427 +0000 UTC m=+400.621381693" Nov 28 14:36:19 crc kubenswrapper[4817]: I1128 14:36:19.013148 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7854b76c6-rhmh8" Nov 28 14:36:19 crc kubenswrapper[4817]: I1128 14:36:19.746471 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b466dce1-0ab0-4467-98d9-8f760ee011df" path="/var/lib/kubelet/pods/b466dce1-0ab0-4467-98d9-8f760ee011df/volumes" Nov 28 14:36:24 crc kubenswrapper[4817]: I1128 14:36:24.045118 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:36:24 crc kubenswrapper[4817]: I1128 14:36:24.045208 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:36:34 crc kubenswrapper[4817]: I1128 14:36:34.717811 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" podUID="f0ef5080-e168-4e13-92d5-8a41f51610da" containerName="registry" containerID="cri-o://2b88bf25abf875fba580d01c5232e33f33f772f716e39877873af13b367f89a7" gracePeriod=30 Nov 28 14:36:35 crc kubenswrapper[4817]: I1128 14:36:35.107756 4817 generic.go:334] "Generic (PLEG): container finished" podID="f0ef5080-e168-4e13-92d5-8a41f51610da" containerID="2b88bf25abf875fba580d01c5232e33f33f772f716e39877873af13b367f89a7" exitCode=0 Nov 28 14:36:35 crc kubenswrapper[4817]: I1128 14:36:35.107825 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" event={"ID":"f0ef5080-e168-4e13-92d5-8a41f51610da","Type":"ContainerDied","Data":"2b88bf25abf875fba580d01c5232e33f33f772f716e39877873af13b367f89a7"} Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.027307 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.115843 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" event={"ID":"f0ef5080-e168-4e13-92d5-8a41f51610da","Type":"ContainerDied","Data":"d89754a5ab79e371c526aeb9967fe8683b3b5ed7dc7b5349241d1722303f4c0d"} Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.115912 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v76hp" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.115929 4817 scope.go:117] "RemoveContainer" containerID="2b88bf25abf875fba580d01c5232e33f33f772f716e39877873af13b367f89a7" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142010 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-certificates\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142096 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f0ef5080-e168-4e13-92d5-8a41f51610da-ca-trust-extracted\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142151 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f0ef5080-e168-4e13-92d5-8a41f51610da-installation-pull-secrets\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142215 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-trusted-ca\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142257 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m255s\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-kube-api-access-m255s\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142293 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-bound-sa-token\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142455 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.142523 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-tls\") pod \"f0ef5080-e168-4e13-92d5-8a41f51610da\" (UID: \"f0ef5080-e168-4e13-92d5-8a41f51610da\") " Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.143547 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.143658 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.161424 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.161477 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ef5080-e168-4e13-92d5-8a41f51610da-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.161502 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.161544 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-kube-api-access-m255s" (OuterVolumeSpecName: "kube-api-access-m255s") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "kube-api-access-m255s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.161701 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.166685 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0ef5080-e168-4e13-92d5-8a41f51610da-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f0ef5080-e168-4e13-92d5-8a41f51610da" (UID: "f0ef5080-e168-4e13-92d5-8a41f51610da"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244509 4817 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244569 4817 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244592 4817 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f0ef5080-e168-4e13-92d5-8a41f51610da-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244612 4817 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f0ef5080-e168-4e13-92d5-8a41f51610da-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244629 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0ef5080-e168-4e13-92d5-8a41f51610da-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244648 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m255s\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-kube-api-access-m255s\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.244665 4817 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0ef5080-e168-4e13-92d5-8a41f51610da-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.465961 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v76hp"] Nov 28 14:36:36 crc kubenswrapper[4817]: I1128 14:36:36.475069 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v76hp"] Nov 28 14:36:37 crc kubenswrapper[4817]: I1128 14:36:37.746017 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0ef5080-e168-4e13-92d5-8a41f51610da" path="/var/lib/kubelet/pods/f0ef5080-e168-4e13-92d5-8a41f51610da/volumes" Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.044799 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.045438 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.045494 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.046117 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d55fccdcaeded99ec2f598c53c1bf31519abf005e37cbd34c5557b27ee85582"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.046184 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://7d55fccdcaeded99ec2f598c53c1bf31519abf005e37cbd34c5557b27ee85582" gracePeriod=600 Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.235959 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="7d55fccdcaeded99ec2f598c53c1bf31519abf005e37cbd34c5557b27ee85582" exitCode=0 Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.236054 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"7d55fccdcaeded99ec2f598c53c1bf31519abf005e37cbd34c5557b27ee85582"} Nov 28 14:36:54 crc kubenswrapper[4817]: I1128 14:36:54.236376 4817 scope.go:117] "RemoveContainer" containerID="4e05251d8a1be61759e2e9dec613276b8b011756bef10b199cceb51833add08a" Nov 28 14:36:55 crc kubenswrapper[4817]: I1128 14:36:55.246922 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"93f0ad5fc8ef7c51ce6b53865c3f234a722e3cdac3537164a439b62ef13ee382"} Nov 28 14:38:37 crc kubenswrapper[4817]: I1128 14:38:37.992441 4817 scope.go:117] "RemoveContainer" containerID="1d068147cbbf0344fc9f546245c4f2fbc1bcc6db806a51b5fa0f7721ca0d40a9" Nov 28 14:38:54 crc kubenswrapper[4817]: I1128 14:38:54.045762 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:38:54 crc kubenswrapper[4817]: I1128 14:38:54.046374 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.227941 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-fmdv9"] Nov 28 14:39:08 crc kubenswrapper[4817]: E1128 14:39:08.229217 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ef5080-e168-4e13-92d5-8a41f51610da" containerName="registry" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.229231 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ef5080-e168-4e13-92d5-8a41f51610da" containerName="registry" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.229402 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ef5080-e168-4e13-92d5-8a41f51610da" containerName="registry" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.229943 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.231992 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.232035 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.233465 4817 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-spz7m" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.246612 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-fmdv9"] Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.247519 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d2rzq"] Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.250769 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d2rzq" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.252673 4817 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bffvp" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.262284 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xfd76"] Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.263004 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.268196 4817 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-bmrsb" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.273606 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xfd76"] Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.278761 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d2rzq"] Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.378955 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm5m6\" (UniqueName: \"kubernetes.io/projected/d33ac74c-861f-4b72-8ba8-38e14244b510-kube-api-access-vm5m6\") pod \"cert-manager-5b446d88c5-d2rzq\" (UID: \"d33ac74c-861f-4b72-8ba8-38e14244b510\") " pod="cert-manager/cert-manager-5b446d88c5-d2rzq" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.379351 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw79b\" (UniqueName: \"kubernetes.io/projected/aee642f7-2003-406e-b5fd-b5a9b6c6e1fe-kube-api-access-lw79b\") pod \"cert-manager-webhook-5655c58dd6-xfd76\" (UID: \"aee642f7-2003-406e-b5fd-b5a9b6c6e1fe\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.379494 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stt8c\" (UniqueName: \"kubernetes.io/projected/d6857358-29dc-4bfa-89cc-503c7d452236-kube-api-access-stt8c\") pod \"cert-manager-cainjector-7f985d654d-fmdv9\" (UID: \"d6857358-29dc-4bfa-89cc-503c7d452236\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.480778 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm5m6\" (UniqueName: \"kubernetes.io/projected/d33ac74c-861f-4b72-8ba8-38e14244b510-kube-api-access-vm5m6\") pod \"cert-manager-5b446d88c5-d2rzq\" (UID: \"d33ac74c-861f-4b72-8ba8-38e14244b510\") " pod="cert-manager/cert-manager-5b446d88c5-d2rzq" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.480850 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw79b\" (UniqueName: \"kubernetes.io/projected/aee642f7-2003-406e-b5fd-b5a9b6c6e1fe-kube-api-access-lw79b\") pod \"cert-manager-webhook-5655c58dd6-xfd76\" (UID: \"aee642f7-2003-406e-b5fd-b5a9b6c6e1fe\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.480934 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stt8c\" (UniqueName: \"kubernetes.io/projected/d6857358-29dc-4bfa-89cc-503c7d452236-kube-api-access-stt8c\") pod \"cert-manager-cainjector-7f985d654d-fmdv9\" (UID: \"d6857358-29dc-4bfa-89cc-503c7d452236\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.501950 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw79b\" (UniqueName: \"kubernetes.io/projected/aee642f7-2003-406e-b5fd-b5a9b6c6e1fe-kube-api-access-lw79b\") pod \"cert-manager-webhook-5655c58dd6-xfd76\" (UID: \"aee642f7-2003-406e-b5fd-b5a9b6c6e1fe\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.504315 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stt8c\" (UniqueName: \"kubernetes.io/projected/d6857358-29dc-4bfa-89cc-503c7d452236-kube-api-access-stt8c\") pod \"cert-manager-cainjector-7f985d654d-fmdv9\" (UID: \"d6857358-29dc-4bfa-89cc-503c7d452236\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.504891 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm5m6\" (UniqueName: \"kubernetes.io/projected/d33ac74c-861f-4b72-8ba8-38e14244b510-kube-api-access-vm5m6\") pod \"cert-manager-5b446d88c5-d2rzq\" (UID: \"d33ac74c-861f-4b72-8ba8-38e14244b510\") " pod="cert-manager/cert-manager-5b446d88c5-d2rzq" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.546935 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.574650 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d2rzq" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.584196 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.782075 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-fmdv9"] Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.793611 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.814887 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d2rzq"] Nov 28 14:39:08 crc kubenswrapper[4817]: W1128 14:39:08.821345 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd33ac74c_861f_4b72_8ba8_38e14244b510.slice/crio-dec3cd36b333fc2d53f4ac5cba9b4808cfb1d67e583ee6705c92111facbf97c9 WatchSource:0}: Error finding container dec3cd36b333fc2d53f4ac5cba9b4808cfb1d67e583ee6705c92111facbf97c9: Status 404 returned error can't find the container with id dec3cd36b333fc2d53f4ac5cba9b4808cfb1d67e583ee6705c92111facbf97c9 Nov 28 14:39:08 crc kubenswrapper[4817]: I1128 14:39:08.856681 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xfd76"] Nov 28 14:39:08 crc kubenswrapper[4817]: W1128 14:39:08.861587 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaee642f7_2003_406e_b5fd_b5a9b6c6e1fe.slice/crio-9ba8dc5850409f5b9fdb22612735c78a624ca7a43e5eb2312a00007ad58afe27 WatchSource:0}: Error finding container 9ba8dc5850409f5b9fdb22612735c78a624ca7a43e5eb2312a00007ad58afe27: Status 404 returned error can't find the container with id 9ba8dc5850409f5b9fdb22612735c78a624ca7a43e5eb2312a00007ad58afe27 Nov 28 14:39:09 crc kubenswrapper[4817]: I1128 14:39:09.109609 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" event={"ID":"aee642f7-2003-406e-b5fd-b5a9b6c6e1fe","Type":"ContainerStarted","Data":"9ba8dc5850409f5b9fdb22612735c78a624ca7a43e5eb2312a00007ad58afe27"} Nov 28 14:39:09 crc kubenswrapper[4817]: I1128 14:39:09.112469 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d2rzq" event={"ID":"d33ac74c-861f-4b72-8ba8-38e14244b510","Type":"ContainerStarted","Data":"dec3cd36b333fc2d53f4ac5cba9b4808cfb1d67e583ee6705c92111facbf97c9"} Nov 28 14:39:09 crc kubenswrapper[4817]: I1128 14:39:09.113910 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" event={"ID":"d6857358-29dc-4bfa-89cc-503c7d452236","Type":"ContainerStarted","Data":"42a51631054dfa51ffb35179ce9aed551a114a85e2a88ab583e2fdc5de8de55e"} Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.138360 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" event={"ID":"aee642f7-2003-406e-b5fd-b5a9b6c6e1fe","Type":"ContainerStarted","Data":"5456d6cae4e55772849aa1e16a23991e7fe28ad840da9ce7587ed8960ad2157c"} Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.138920 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.140476 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d2rzq" event={"ID":"d33ac74c-861f-4b72-8ba8-38e14244b510","Type":"ContainerStarted","Data":"e5f3409439083f7a01bb137193e66b8d4e40b3914220be379c4c4b9439afd409"} Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.141817 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" event={"ID":"d6857358-29dc-4bfa-89cc-503c7d452236","Type":"ContainerStarted","Data":"7e0dad7f55a59f8afdcd453960080463c9d8a510ef32ce88affa198d56cca4f7"} Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.159095 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" podStartSLOduration=1.545427075 podStartE2EDuration="5.159072316s" podCreationTimestamp="2025-11-28 14:39:08 +0000 UTC" firstStartedPulling="2025-11-28 14:39:08.863372916 +0000 UTC m=+571.451351182" lastFinishedPulling="2025-11-28 14:39:12.477018157 +0000 UTC m=+575.064996423" observedRunningTime="2025-11-28 14:39:13.155944489 +0000 UTC m=+575.743922795" watchObservedRunningTime="2025-11-28 14:39:13.159072316 +0000 UTC m=+575.747050582" Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.171262 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-d2rzq" podStartSLOduration=1.520684667 podStartE2EDuration="5.171240035s" podCreationTimestamp="2025-11-28 14:39:08 +0000 UTC" firstStartedPulling="2025-11-28 14:39:08.827294029 +0000 UTC m=+571.415272305" lastFinishedPulling="2025-11-28 14:39:12.477849407 +0000 UTC m=+575.065827673" observedRunningTime="2025-11-28 14:39:13.168340674 +0000 UTC m=+575.756318950" watchObservedRunningTime="2025-11-28 14:39:13.171240035 +0000 UTC m=+575.759218301" Nov 28 14:39:13 crc kubenswrapper[4817]: I1128 14:39:13.194422 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-fmdv9" podStartSLOduration=1.338104126 podStartE2EDuration="5.194401425s" podCreationTimestamp="2025-11-28 14:39:08 +0000 UTC" firstStartedPulling="2025-11-28 14:39:08.793327613 +0000 UTC m=+571.381305879" lastFinishedPulling="2025-11-28 14:39:12.649624892 +0000 UTC m=+575.237603178" observedRunningTime="2025-11-28 14:39:13.192912178 +0000 UTC m=+575.780890444" watchObservedRunningTime="2025-11-28 14:39:13.194401425 +0000 UTC m=+575.782379691" Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.185313 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bktf8"] Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.186448 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-controller" containerID="cri-o://fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.186947 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="sbdb" containerID="cri-o://8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.187047 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="nbdb" containerID="cri-o://50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.187091 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="northd" containerID="cri-o://86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.187151 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.187190 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-node" containerID="cri-o://d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.187226 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-acl-logging" containerID="cri-o://2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.262237 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" containerID="cri-o://0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" gracePeriod=30 Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.552633 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 is running failed: container process not found" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.553061 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 is running failed: container process not found" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.553331 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 is running failed: container process not found" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.553408 4817 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="nbdb" Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.554460 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.555585 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.556768 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Nov 28 14:39:18 crc kubenswrapper[4817]: E1128 14:39:18.556837 4817 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="sbdb" Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.587311 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xfd76" Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.912849 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/3.log" Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.919342 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovn-acl-logging/0.log" Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.923060 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovn-controller/0.log" Nov 28 14:39:18 crc kubenswrapper[4817]: I1128 14:39:18.925454 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.017009 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9fnhn"] Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.017761 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.017945 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.018086 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-acl-logging" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.018224 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-acl-logging" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.018394 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.018534 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.018666 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="northd" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.018853 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="northd" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.019033 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.019200 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.019357 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="sbdb" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.019496 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="sbdb" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.019608 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.019696 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.019823 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.019913 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.020010 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.020099 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.020185 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kubecfg-setup" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.020302 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kubecfg-setup" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.020393 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="nbdb" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.020480 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="nbdb" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.020566 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-node" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.020661 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-node" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.020949 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="nbdb" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.021065 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.021157 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.021250 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.021344 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-acl-logging" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.021441 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.021541 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="northd" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.022199 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovn-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.022332 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.022522 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="sbdb" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.022689 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="kube-rbac-proxy-node" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.023217 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.023399 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.023847 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerName="ovnkube-controller" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.027605 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124004 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-config\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124098 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovn-node-metrics-cert\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124134 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-ovn-kubernetes\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124164 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-openvswitch\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124212 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-log-socket\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124254 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-bin\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124317 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-kubelet\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124404 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124984 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.124997 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-log-socket" (OuterVolumeSpecName: "log-socket") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.125087 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.125130 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.125173 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.125852 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.125015 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-script-lib\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.126560 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.126457 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-netd\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.127022 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-ovn\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.127404 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-systemd-units\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.127705 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-node-log\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.128069 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvq6d\" (UniqueName: \"kubernetes.io/projected/8fa7491e-b213-4068-b2a8-77bf4979fc38-kube-api-access-qvq6d\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.127253 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.127485 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.127812 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-node-log" (OuterVolumeSpecName: "node-log") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.128547 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.129078 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-netns\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.129341 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.129598 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.129973 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-env-overrides\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.130318 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.130413 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.130215 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-etc-openvswitch\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.131026 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-systemd\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.131939 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-slash\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.132179 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-var-lib-openvswitch\") pod \"8fa7491e-b213-4068-b2a8-77bf4979fc38\" (UID: \"8fa7491e-b213-4068-b2a8-77bf4979fc38\") " Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.132067 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-slash" (OuterVolumeSpecName: "host-slash") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.132311 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.133112 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-cni-bin\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.133330 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-run-ovn-kubernetes\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.133507 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-ovnkube-config\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.133683 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/741e36c7-abed-40b1-92c6-12922b01f50e-ovn-node-metrics-cert\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.133929 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-etc-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.134118 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-ovnkube-script-lib\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.134303 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq7w8\" (UniqueName: \"kubernetes.io/projected/741e36c7-abed-40b1-92c6-12922b01f50e-kube-api-access-wq7w8\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.134551 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-var-lib-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.134745 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-log-socket\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.135123 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.133819 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa7491e-b213-4068-b2a8-77bf4979fc38-kube-api-access-qvq6d" (OuterVolumeSpecName: "kube-api-access-qvq6d") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "kube-api-access-qvq6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.134178 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.135779 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-run-netns\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.135976 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-env-overrides\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.136170 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-cni-netd\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.136329 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.136496 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-node-log\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.136668 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-ovn\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.136859 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-systemd-units\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137023 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-systemd\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137160 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-slash\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137358 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-kubelet\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137552 4817 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137678 4817 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137816 4817 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.137944 4817 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138066 4817 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-node-log\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138307 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvq6d\" (UniqueName: \"kubernetes.io/projected/8fa7491e-b213-4068-b2a8-77bf4979fc38-kube-api-access-qvq6d\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138426 4817 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138541 4817 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138647 4817 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138818 4817 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.138976 4817 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-slash\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139086 4817 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139198 4817 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139310 4817 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8fa7491e-b213-4068-b2a8-77bf4979fc38-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139423 4817 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139537 4817 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139652 4817 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-log-socket\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139791 4817 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.139917 4817 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.144250 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "8fa7491e-b213-4068-b2a8-77bf4979fc38" (UID: "8fa7491e-b213-4068-b2a8-77bf4979fc38"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.185108 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/2.log" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.186044 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/1.log" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.186127 4817 generic.go:334] "Generic (PLEG): container finished" podID="89086178-4127-42a3-8454-68e6d00c912a" containerID="c798287a6a545dc85c7121d38a6303407847b11a9390f3470f4f8f465b90166a" exitCode=2 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.186227 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerDied","Data":"c798287a6a545dc85c7121d38a6303407847b11a9390f3470f4f8f465b90166a"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.186275 4817 scope.go:117] "RemoveContainer" containerID="58471f5ffca19830d96537ddad9aa427d05c26c1f49276b5f337238f891f9ecf" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.187696 4817 scope.go:117] "RemoveContainer" containerID="c798287a6a545dc85c7121d38a6303407847b11a9390f3470f4f8f465b90166a" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.188325 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-sbhgx_openshift-multus(89086178-4127-42a3-8454-68e6d00c912a)\"" pod="openshift-multus/multus-sbhgx" podUID="89086178-4127-42a3-8454-68e6d00c912a" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.194291 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovnkube-controller/3.log" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.204912 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovn-acl-logging/0.log" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206186 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bktf8_8fa7491e-b213-4068-b2a8-77bf4979fc38/ovn-controller/0.log" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206783 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" exitCode=0 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206826 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" exitCode=0 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206835 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" exitCode=0 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206842 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" exitCode=0 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206896 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" exitCode=0 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206903 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" exitCode=0 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206909 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" exitCode=143 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206917 4817 generic.go:334] "Generic (PLEG): container finished" podID="8fa7491e-b213-4068-b2a8-77bf4979fc38" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" exitCode=143 Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206939 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206964 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206975 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206985 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.206993 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207002 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207012 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207016 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207023 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207830 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207890 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207898 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207904 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207909 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207915 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207921 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207977 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.207991 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208002 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208008 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208058 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208065 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208071 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208076 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208120 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208127 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208132 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208137 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208163 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208173 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208179 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208204 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208211 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208217 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208224 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208230 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208237 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208243 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208250 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208283 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bktf8" event={"ID":"8fa7491e-b213-4068-b2a8-77bf4979fc38","Type":"ContainerDied","Data":"07f7ea85e9aafc4c3165b2709de8c98e364135b61299b942e1516da44313ad77"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208294 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208302 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208308 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208314 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208319 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208324 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208329 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208335 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208339 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.208370 4817 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.240901 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-etc-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.240967 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-ovnkube-script-lib\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241034 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq7w8\" (UniqueName: \"kubernetes.io/projected/741e36c7-abed-40b1-92c6-12922b01f50e-kube-api-access-wq7w8\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241055 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-etc-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241105 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-var-lib-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241066 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-var-lib-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241207 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-log-socket\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241244 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241279 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-run-netns\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241297 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-env-overrides\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241323 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-cni-netd\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241341 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241371 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-node-log\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241408 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-ovn\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241425 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-systemd-units\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241446 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-systemd\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241459 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-slash\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241522 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-kubelet\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241551 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-cni-bin\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241567 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-run-ovn-kubernetes\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241581 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-ovnkube-config\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241600 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/741e36c7-abed-40b1-92c6-12922b01f50e-ovn-node-metrics-cert\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.241661 4817 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8fa7491e-b213-4068-b2a8-77bf4979fc38-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242112 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-ovnkube-script-lib\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242264 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-slash\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242307 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242278 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-log-socket\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242347 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-run-netns\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242380 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-openvswitch\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242383 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-cni-netd\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242383 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-ovn\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242410 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-node-log\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242420 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-systemd-units\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242439 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-cni-bin\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242468 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-run-systemd\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242471 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-kubelet\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242494 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/741e36c7-abed-40b1-92c6-12922b01f50e-host-run-ovn-kubernetes\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.242919 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-env-overrides\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.243553 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/741e36c7-abed-40b1-92c6-12922b01f50e-ovnkube-config\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.250422 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/741e36c7-abed-40b1-92c6-12922b01f50e-ovn-node-metrics-cert\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.252714 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bktf8"] Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.257820 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bktf8"] Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.268589 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq7w8\" (UniqueName: \"kubernetes.io/projected/741e36c7-abed-40b1-92c6-12922b01f50e-kube-api-access-wq7w8\") pod \"ovnkube-node-9fnhn\" (UID: \"741e36c7-abed-40b1-92c6-12922b01f50e\") " pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.346095 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.368379 4817 scope.go:117] "RemoveContainer" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.388881 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.454584 4817 scope.go:117] "RemoveContainer" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.480483 4817 scope.go:117] "RemoveContainer" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.506676 4817 scope.go:117] "RemoveContainer" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.526805 4817 scope.go:117] "RemoveContainer" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.547502 4817 scope.go:117] "RemoveContainer" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.595136 4817 scope.go:117] "RemoveContainer" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.611662 4817 scope.go:117] "RemoveContainer" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.629888 4817 scope.go:117] "RemoveContainer" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644030 4817 scope.go:117] "RemoveContainer" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.644329 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": container with ID starting with 0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543 not found: ID does not exist" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644364 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} err="failed to get container status \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": rpc error: code = NotFound desc = could not find container \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": container with ID starting with 0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644389 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.644618 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": container with ID starting with 8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1 not found: ID does not exist" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644646 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} err="failed to get container status \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": rpc error: code = NotFound desc = could not find container \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": container with ID starting with 8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644664 4817 scope.go:117] "RemoveContainer" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.644908 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": container with ID starting with 8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc not found: ID does not exist" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644937 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} err="failed to get container status \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": rpc error: code = NotFound desc = could not find container \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": container with ID starting with 8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.644956 4817 scope.go:117] "RemoveContainer" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.645314 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": container with ID starting with 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 not found: ID does not exist" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.645372 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} err="failed to get container status \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": rpc error: code = NotFound desc = could not find container \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": container with ID starting with 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.645412 4817 scope.go:117] "RemoveContainer" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.645700 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": container with ID starting with 86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f not found: ID does not exist" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.645762 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} err="failed to get container status \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": rpc error: code = NotFound desc = could not find container \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": container with ID starting with 86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.645787 4817 scope.go:117] "RemoveContainer" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.646046 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": container with ID starting with d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1 not found: ID does not exist" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.646082 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} err="failed to get container status \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": rpc error: code = NotFound desc = could not find container \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": container with ID starting with d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.646104 4817 scope.go:117] "RemoveContainer" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.646345 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": container with ID starting with d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2 not found: ID does not exist" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.646379 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} err="failed to get container status \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": rpc error: code = NotFound desc = could not find container \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": container with ID starting with d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.646401 4817 scope.go:117] "RemoveContainer" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.646674 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": container with ID starting with 2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119 not found: ID does not exist" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.646710 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} err="failed to get container status \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": rpc error: code = NotFound desc = could not find container \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": container with ID starting with 2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.646757 4817 scope.go:117] "RemoveContainer" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.647057 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": container with ID starting with fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9 not found: ID does not exist" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.647092 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} err="failed to get container status \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": rpc error: code = NotFound desc = could not find container \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": container with ID starting with fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.647114 4817 scope.go:117] "RemoveContainer" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" Nov 28 14:39:19 crc kubenswrapper[4817]: E1128 14:39:19.647304 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": container with ID starting with e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c not found: ID does not exist" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.647340 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} err="failed to get container status \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": rpc error: code = NotFound desc = could not find container \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": container with ID starting with e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.647363 4817 scope.go:117] "RemoveContainer" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.647586 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} err="failed to get container status \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": rpc error: code = NotFound desc = could not find container \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": container with ID starting with 0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.647617 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.648907 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} err="failed to get container status \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": rpc error: code = NotFound desc = could not find container \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": container with ID starting with 8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.648941 4817 scope.go:117] "RemoveContainer" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649191 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} err="failed to get container status \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": rpc error: code = NotFound desc = could not find container \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": container with ID starting with 8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649221 4817 scope.go:117] "RemoveContainer" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649457 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} err="failed to get container status \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": rpc error: code = NotFound desc = could not find container \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": container with ID starting with 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649489 4817 scope.go:117] "RemoveContainer" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649701 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} err="failed to get container status \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": rpc error: code = NotFound desc = could not find container \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": container with ID starting with 86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649749 4817 scope.go:117] "RemoveContainer" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.649990 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} err="failed to get container status \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": rpc error: code = NotFound desc = could not find container \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": container with ID starting with d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.650022 4817 scope.go:117] "RemoveContainer" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.650240 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} err="failed to get container status \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": rpc error: code = NotFound desc = could not find container \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": container with ID starting with d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.650269 4817 scope.go:117] "RemoveContainer" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.650767 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} err="failed to get container status \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": rpc error: code = NotFound desc = could not find container \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": container with ID starting with 2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.650799 4817 scope.go:117] "RemoveContainer" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651009 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} err="failed to get container status \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": rpc error: code = NotFound desc = could not find container \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": container with ID starting with fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651038 4817 scope.go:117] "RemoveContainer" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651222 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} err="failed to get container status \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": rpc error: code = NotFound desc = could not find container \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": container with ID starting with e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651253 4817 scope.go:117] "RemoveContainer" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651494 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} err="failed to get container status \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": rpc error: code = NotFound desc = could not find container \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": container with ID starting with 0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651525 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651782 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} err="failed to get container status \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": rpc error: code = NotFound desc = could not find container \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": container with ID starting with 8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.651810 4817 scope.go:117] "RemoveContainer" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652052 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} err="failed to get container status \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": rpc error: code = NotFound desc = could not find container \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": container with ID starting with 8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652083 4817 scope.go:117] "RemoveContainer" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652336 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} err="failed to get container status \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": rpc error: code = NotFound desc = could not find container \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": container with ID starting with 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652366 4817 scope.go:117] "RemoveContainer" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652686 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} err="failed to get container status \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": rpc error: code = NotFound desc = could not find container \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": container with ID starting with 86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652748 4817 scope.go:117] "RemoveContainer" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652966 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} err="failed to get container status \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": rpc error: code = NotFound desc = could not find container \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": container with ID starting with d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.652999 4817 scope.go:117] "RemoveContainer" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.653187 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} err="failed to get container status \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": rpc error: code = NotFound desc = could not find container \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": container with ID starting with d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.653212 4817 scope.go:117] "RemoveContainer" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.653401 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} err="failed to get container status \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": rpc error: code = NotFound desc = could not find container \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": container with ID starting with 2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.653438 4817 scope.go:117] "RemoveContainer" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.653845 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} err="failed to get container status \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": rpc error: code = NotFound desc = could not find container \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": container with ID starting with fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.653866 4817 scope.go:117] "RemoveContainer" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.654133 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} err="failed to get container status \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": rpc error: code = NotFound desc = could not find container \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": container with ID starting with e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.654152 4817 scope.go:117] "RemoveContainer" containerID="0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.654460 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543"} err="failed to get container status \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": rpc error: code = NotFound desc = could not find container \"0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543\": container with ID starting with 0490f8f0b24c4c13589e8449f7d01ead290a024e4a8fada3045df3c4153ac543 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.654497 4817 scope.go:117] "RemoveContainer" containerID="8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.654933 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1"} err="failed to get container status \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": rpc error: code = NotFound desc = could not find container \"8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1\": container with ID starting with 8830f347f1e40338bc6727f4dd175b10e9a4967cef0cc8779e498378aeeb10f1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.654966 4817 scope.go:117] "RemoveContainer" containerID="8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.655295 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc"} err="failed to get container status \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": rpc error: code = NotFound desc = could not find container \"8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc\": container with ID starting with 8eade5c2a78193aaacd25e29df2a60b4b69e095bf9388911e5441d5daa9e83bc not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.655334 4817 scope.go:117] "RemoveContainer" containerID="50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.655756 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0"} err="failed to get container status \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": rpc error: code = NotFound desc = could not find container \"50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0\": container with ID starting with 50720d176306bdbe246dddeb42f2188c29457a289c9f32c7c3735ab243081ab0 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.655790 4817 scope.go:117] "RemoveContainer" containerID="86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.656259 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f"} err="failed to get container status \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": rpc error: code = NotFound desc = could not find container \"86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f\": container with ID starting with 86ec3c99d7599733b23f6fc2929b427b8326b30fdd9a948f0c0ebb107fe9f74f not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.656294 4817 scope.go:117] "RemoveContainer" containerID="d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.656575 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1"} err="failed to get container status \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": rpc error: code = NotFound desc = could not find container \"d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1\": container with ID starting with d0607e5f77f31b0af7cc9cef221a64180cfbeaf5f64d0cb50e8d03bca88f56b1 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.656612 4817 scope.go:117] "RemoveContainer" containerID="d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.657041 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2"} err="failed to get container status \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": rpc error: code = NotFound desc = could not find container \"d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2\": container with ID starting with d42b0392560a10d06be96a780722b0bfeecf43438bb6848c2f3cbc3e21744ec2 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.657076 4817 scope.go:117] "RemoveContainer" containerID="2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.657380 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119"} err="failed to get container status \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": rpc error: code = NotFound desc = could not find container \"2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119\": container with ID starting with 2a078abeab099f46e3bf0c7da7a9ce6cf050cb965a7c735c0c747c2eafeac119 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.657412 4817 scope.go:117] "RemoveContainer" containerID="fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.657653 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9"} err="failed to get container status \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": rpc error: code = NotFound desc = could not find container \"fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9\": container with ID starting with fe6b89b48795b4c0915725a9bbfe0ee3baab47fca85c2a17c6479e6bcf1e38a9 not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.657691 4817 scope.go:117] "RemoveContainer" containerID="e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.658092 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c"} err="failed to get container status \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": rpc error: code = NotFound desc = could not find container \"e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c\": container with ID starting with e1c4f70411c5e1e5ad2ab405f96b0fa984d2a2e9bbe7f488fc46d15a0b089c6c not found: ID does not exist" Nov 28 14:39:19 crc kubenswrapper[4817]: I1128 14:39:19.751490 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa7491e-b213-4068-b2a8-77bf4979fc38" path="/var/lib/kubelet/pods/8fa7491e-b213-4068-b2a8-77bf4979fc38/volumes" Nov 28 14:39:20 crc kubenswrapper[4817]: I1128 14:39:20.216695 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/2.log" Nov 28 14:39:20 crc kubenswrapper[4817]: I1128 14:39:20.221084 4817 generic.go:334] "Generic (PLEG): container finished" podID="741e36c7-abed-40b1-92c6-12922b01f50e" containerID="11337a0e0d44611ac6418521c1ace41b85ae60f882a00bd2b0469dad86f4f39e" exitCode=0 Nov 28 14:39:20 crc kubenswrapper[4817]: I1128 14:39:20.221137 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerDied","Data":"11337a0e0d44611ac6418521c1ace41b85ae60f882a00bd2b0469dad86f4f39e"} Nov 28 14:39:20 crc kubenswrapper[4817]: I1128 14:39:20.221186 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"ab73ed350afcea8ae2b50d1c17cb6daedc0bd6b4151ec9f0bc8ce1d13cbc22e7"} Nov 28 14:39:21 crc kubenswrapper[4817]: I1128 14:39:21.232101 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"d7742724b011512d88ec2c7d257c2ff19804dcfb5746ebff4d509aa03d4c48c3"} Nov 28 14:39:21 crc kubenswrapper[4817]: I1128 14:39:21.232415 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"45fad8ebfadaa2b760ee35d87e427b33965578978fb1a7d742ce12fd35cb51a9"} Nov 28 14:39:21 crc kubenswrapper[4817]: I1128 14:39:21.232429 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"bf387cb15ca7a3f2b75628f184bb97041a91c0b9c4d1d14fb842b0a184db9e23"} Nov 28 14:39:21 crc kubenswrapper[4817]: I1128 14:39:21.232443 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"b7a092080b603dcf9ca563cdf9a5deb46d10f423dcd539668cc9cbe81fbf6fe3"} Nov 28 14:39:21 crc kubenswrapper[4817]: I1128 14:39:21.232455 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"3daddaaeaea84d395dda9318ac2b39c70597c732fae29855931fd6aec240bd5d"} Nov 28 14:39:21 crc kubenswrapper[4817]: I1128 14:39:21.232467 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"4fc7ce200ef0276063dc290e95c1aae7eeb8e258c208727fd97543f0db983947"} Nov 28 14:39:24 crc kubenswrapper[4817]: I1128 14:39:24.045752 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:39:24 crc kubenswrapper[4817]: I1128 14:39:24.046162 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:39:24 crc kubenswrapper[4817]: I1128 14:39:24.259652 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"1a0f5809289a76329e7d1e42d30ce7e60f315ab69cc565a875dd312da60e84a3"} Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.276559 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" event={"ID":"741e36c7-abed-40b1-92c6-12922b01f50e","Type":"ContainerStarted","Data":"8ffe680640479f76ad1692840ed78a64325268ab7505e2f5095c7373ef55015c"} Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.277128 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.277144 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.277152 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.302634 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.312433 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:26 crc kubenswrapper[4817]: I1128 14:39:26.326855 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" podStartSLOduration=8.32682974 podStartE2EDuration="8.32682974s" podCreationTimestamp="2025-11-28 14:39:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:39:26.322559514 +0000 UTC m=+588.910537810" watchObservedRunningTime="2025-11-28 14:39:26.32682974 +0000 UTC m=+588.914808006" Nov 28 14:39:30 crc kubenswrapper[4817]: I1128 14:39:30.737415 4817 scope.go:117] "RemoveContainer" containerID="c798287a6a545dc85c7121d38a6303407847b11a9390f3470f4f8f465b90166a" Nov 28 14:39:30 crc kubenswrapper[4817]: E1128 14:39:30.739015 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-sbhgx_openshift-multus(89086178-4127-42a3-8454-68e6d00c912a)\"" pod="openshift-multus/multus-sbhgx" podUID="89086178-4127-42a3-8454-68e6d00c912a" Nov 28 14:39:41 crc kubenswrapper[4817]: I1128 14:39:41.737364 4817 scope.go:117] "RemoveContainer" containerID="c798287a6a545dc85c7121d38a6303407847b11a9390f3470f4f8f465b90166a" Nov 28 14:39:42 crc kubenswrapper[4817]: I1128 14:39:42.383284 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sbhgx_89086178-4127-42a3-8454-68e6d00c912a/kube-multus/2.log" Nov 28 14:39:42 crc kubenswrapper[4817]: I1128 14:39:42.383667 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sbhgx" event={"ID":"89086178-4127-42a3-8454-68e6d00c912a","Type":"ContainerStarted","Data":"d164b86aa764c73774986833f9c40fc23254594d9a48635225877ba2e739e97e"} Nov 28 14:39:49 crc kubenswrapper[4817]: I1128 14:39:49.375730 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9fnhn" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.044952 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.045337 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.045390 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.046040 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93f0ad5fc8ef7c51ce6b53865c3f234a722e3cdac3537164a439b62ef13ee382"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.046107 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://93f0ad5fc8ef7c51ce6b53865c3f234a722e3cdac3537164a439b62ef13ee382" gracePeriod=600 Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.130818 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt"] Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.132152 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.134469 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.140206 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt"] Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.296147 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.296211 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdkp4\" (UniqueName: \"kubernetes.io/projected/83c91d8d-d9d8-4952-ba8b-b57f1d003806-kube-api-access-rdkp4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.296257 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.397845 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.397939 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdkp4\" (UniqueName: \"kubernetes.io/projected/83c91d8d-d9d8-4952-ba8b-b57f1d003806-kube-api-access-rdkp4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.398029 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.398932 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.399235 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.418243 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdkp4\" (UniqueName: \"kubernetes.io/projected/83c91d8d-d9d8-4952-ba8b-b57f1d003806-kube-api-access-rdkp4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.456421 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:39:54 crc kubenswrapper[4817]: I1128 14:39:54.681973 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt"] Nov 28 14:39:55 crc kubenswrapper[4817]: I1128 14:39:55.478101 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" event={"ID":"83c91d8d-d9d8-4952-ba8b-b57f1d003806","Type":"ContainerStarted","Data":"203e314bb9dd8b0bc28363ffd6ad483a35b43edd5cec7d1fbbbdb658da577fb5"} Nov 28 14:39:56 crc kubenswrapper[4817]: I1128 14:39:56.484358 4817 generic.go:334] "Generic (PLEG): container finished" podID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerID="6ba3d0bb8d2dc2f9c9fc165915c9c08017159b9a49532a2b89a1239c15179652" exitCode=0 Nov 28 14:39:56 crc kubenswrapper[4817]: I1128 14:39:56.484402 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" event={"ID":"83c91d8d-d9d8-4952-ba8b-b57f1d003806","Type":"ContainerDied","Data":"6ba3d0bb8d2dc2f9c9fc165915c9c08017159b9a49532a2b89a1239c15179652"} Nov 28 14:39:56 crc kubenswrapper[4817]: I1128 14:39:56.486595 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="93f0ad5fc8ef7c51ce6b53865c3f234a722e3cdac3537164a439b62ef13ee382" exitCode=0 Nov 28 14:39:56 crc kubenswrapper[4817]: I1128 14:39:56.486621 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"93f0ad5fc8ef7c51ce6b53865c3f234a722e3cdac3537164a439b62ef13ee382"} Nov 28 14:39:56 crc kubenswrapper[4817]: I1128 14:39:56.486647 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"e638df1b21b62724d5463d8dfc0e4379271e7a660caed8cd6069eeac89d71b58"} Nov 28 14:39:56 crc kubenswrapper[4817]: I1128 14:39:56.486662 4817 scope.go:117] "RemoveContainer" containerID="7d55fccdcaeded99ec2f598c53c1bf31519abf005e37cbd34c5557b27ee85582" Nov 28 14:39:58 crc kubenswrapper[4817]: I1128 14:39:58.505199 4817 generic.go:334] "Generic (PLEG): container finished" podID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerID="4c3ab43efb5adee6e471512e12c3bd64ec07e2adaee5d99897d0c1edf4016c96" exitCode=0 Nov 28 14:39:58 crc kubenswrapper[4817]: I1128 14:39:58.505290 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" event={"ID":"83c91d8d-d9d8-4952-ba8b-b57f1d003806","Type":"ContainerDied","Data":"4c3ab43efb5adee6e471512e12c3bd64ec07e2adaee5d99897d0c1edf4016c96"} Nov 28 14:39:59 crc kubenswrapper[4817]: I1128 14:39:59.515499 4817 generic.go:334] "Generic (PLEG): container finished" podID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerID="11976ac9f0f4586425b23492ba14e26a90ed57859c8571636e30dd49ae5959cb" exitCode=0 Nov 28 14:39:59 crc kubenswrapper[4817]: I1128 14:39:59.516267 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" event={"ID":"83c91d8d-d9d8-4952-ba8b-b57f1d003806","Type":"ContainerDied","Data":"11976ac9f0f4586425b23492ba14e26a90ed57859c8571636e30dd49ae5959cb"} Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.887825 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.979298 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-bundle\") pod \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.979406 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-util\") pod \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.979455 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdkp4\" (UniqueName: \"kubernetes.io/projected/83c91d8d-d9d8-4952-ba8b-b57f1d003806-kube-api-access-rdkp4\") pod \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\" (UID: \"83c91d8d-d9d8-4952-ba8b-b57f1d003806\") " Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.980578 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-bundle" (OuterVolumeSpecName: "bundle") pod "83c91d8d-d9d8-4952-ba8b-b57f1d003806" (UID: "83c91d8d-d9d8-4952-ba8b-b57f1d003806"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.989021 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c91d8d-d9d8-4952-ba8b-b57f1d003806-kube-api-access-rdkp4" (OuterVolumeSpecName: "kube-api-access-rdkp4") pod "83c91d8d-d9d8-4952-ba8b-b57f1d003806" (UID: "83c91d8d-d9d8-4952-ba8b-b57f1d003806"). InnerVolumeSpecName "kube-api-access-rdkp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:40:00 crc kubenswrapper[4817]: I1128 14:40:00.998879 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-util" (OuterVolumeSpecName: "util") pod "83c91d8d-d9d8-4952-ba8b-b57f1d003806" (UID: "83c91d8d-d9d8-4952-ba8b-b57f1d003806"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:40:01 crc kubenswrapper[4817]: I1128 14:40:01.081051 4817 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:01 crc kubenswrapper[4817]: I1128 14:40:01.081101 4817 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/83c91d8d-d9d8-4952-ba8b-b57f1d003806-util\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:01 crc kubenswrapper[4817]: I1128 14:40:01.081121 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdkp4\" (UniqueName: \"kubernetes.io/projected/83c91d8d-d9d8-4952-ba8b-b57f1d003806-kube-api-access-rdkp4\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:01 crc kubenswrapper[4817]: I1128 14:40:01.534092 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" event={"ID":"83c91d8d-d9d8-4952-ba8b-b57f1d003806","Type":"ContainerDied","Data":"203e314bb9dd8b0bc28363ffd6ad483a35b43edd5cec7d1fbbbdb658da577fb5"} Nov 28 14:40:01 crc kubenswrapper[4817]: I1128 14:40:01.534193 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt" Nov 28 14:40:01 crc kubenswrapper[4817]: I1128 14:40:01.534159 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="203e314bb9dd8b0bc28363ffd6ad483a35b43edd5cec7d1fbbbdb658da577fb5" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.645405 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f"] Nov 28 14:40:05 crc kubenswrapper[4817]: E1128 14:40:05.645890 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="extract" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.645902 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="extract" Nov 28 14:40:05 crc kubenswrapper[4817]: E1128 14:40:05.645912 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="util" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.645918 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="util" Nov 28 14:40:05 crc kubenswrapper[4817]: E1128 14:40:05.645932 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="pull" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.645938 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="pull" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.646029 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c91d8d-d9d8-4952-ba8b-b57f1d003806" containerName="extract" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.646375 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.649504 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.649614 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-9tmps" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.649983 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.664696 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f"] Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.742792 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5564h\" (UniqueName: \"kubernetes.io/projected/68d8ec1c-4246-402e-8da4-53dfe24edbca-kube-api-access-5564h\") pod \"nmstate-operator-5b5b58f5c8-b4g4f\" (UID: \"68d8ec1c-4246-402e-8da4-53dfe24edbca\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.844188 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5564h\" (UniqueName: \"kubernetes.io/projected/68d8ec1c-4246-402e-8da4-53dfe24edbca-kube-api-access-5564h\") pod \"nmstate-operator-5b5b58f5c8-b4g4f\" (UID: \"68d8ec1c-4246-402e-8da4-53dfe24edbca\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.869287 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5564h\" (UniqueName: \"kubernetes.io/projected/68d8ec1c-4246-402e-8da4-53dfe24edbca-kube-api-access-5564h\") pod \"nmstate-operator-5b5b58f5c8-b4g4f\" (UID: \"68d8ec1c-4246-402e-8da4-53dfe24edbca\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" Nov 28 14:40:05 crc kubenswrapper[4817]: I1128 14:40:05.963289 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" Nov 28 14:40:06 crc kubenswrapper[4817]: I1128 14:40:06.162306 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f"] Nov 28 14:40:06 crc kubenswrapper[4817]: I1128 14:40:06.578664 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" event={"ID":"68d8ec1c-4246-402e-8da4-53dfe24edbca","Type":"ContainerStarted","Data":"29cedbe87f8fdb43e1f0ffb427203aca298b13ac07d58bd2b1ad20be8f855156"} Nov 28 14:40:09 crc kubenswrapper[4817]: I1128 14:40:09.601960 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" event={"ID":"68d8ec1c-4246-402e-8da4-53dfe24edbca","Type":"ContainerStarted","Data":"1da7bac7b7d4f2875c4d34b7717f3c0536dfbdb96f002231e3a5b36ad946ec62"} Nov 28 14:40:09 crc kubenswrapper[4817]: I1128 14:40:09.622553 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-b4g4f" podStartSLOduration=2.366904951 podStartE2EDuration="4.622537784s" podCreationTimestamp="2025-11-28 14:40:05 +0000 UTC" firstStartedPulling="2025-11-28 14:40:06.180148201 +0000 UTC m=+628.768126467" lastFinishedPulling="2025-11-28 14:40:08.435781034 +0000 UTC m=+631.023759300" observedRunningTime="2025-11-28 14:40:09.620266068 +0000 UTC m=+632.208244384" watchObservedRunningTime="2025-11-28 14:40:09.622537784 +0000 UTC m=+632.210516050" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.558000 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.560919 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.566346 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.567323 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.567987 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-qhd5x" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.568941 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.575440 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.604157 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kgbh6"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.604925 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.614530 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654433 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-ovs-socket\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654487 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-dbus-socket\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654542 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhr9c\" (UniqueName: \"kubernetes.io/projected/90fcff6a-9f08-479c-8e5b-2745700bc8f1-kube-api-access-rhr9c\") pod \"nmstate-metrics-7f946cbc9-wgrqq\" (UID: \"90fcff6a-9f08-479c-8e5b-2745700bc8f1\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654573 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdftm\" (UniqueName: \"kubernetes.io/projected/8d2164f8-732d-4650-b92a-3fb677b51c25-kube-api-access-wdftm\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654607 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8d2164f8-732d-4650-b92a-3fb677b51c25-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654652 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-nmstate-lock\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.654683 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npmks\" (UniqueName: \"kubernetes.io/projected/feed3d1f-a072-4ccc-b47a-3a5b48b44177-kube-api-access-npmks\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.683154 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.683876 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.686161 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.686169 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.686949 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7jnjb" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.693990 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.755957 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8d2164f8-732d-4650-b92a-3fb677b51c25-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756011 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756060 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc2rf\" (UniqueName: \"kubernetes.io/projected/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-kube-api-access-fc2rf\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756089 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-nmstate-lock\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: E1128 14:40:14.756126 4817 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756151 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npmks\" (UniqueName: \"kubernetes.io/projected/feed3d1f-a072-4ccc-b47a-3a5b48b44177-kube-api-access-npmks\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: E1128 14:40:14.756185 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d2164f8-732d-4650-b92a-3fb677b51c25-tls-key-pair podName:8d2164f8-732d-4650-b92a-3fb677b51c25 nodeName:}" failed. No retries permitted until 2025-11-28 14:40:15.25616898 +0000 UTC m=+637.844147246 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/8d2164f8-732d-4650-b92a-3fb677b51c25-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-pm78n" (UID: "8d2164f8-732d-4650-b92a-3fb677b51c25") : secret "openshift-nmstate-webhook" not found Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756204 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-nmstate-lock\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756215 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-ovs-socket\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756245 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-ovs-socket\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756297 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-dbus-socket\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756349 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhr9c\" (UniqueName: \"kubernetes.io/projected/90fcff6a-9f08-479c-8e5b-2745700bc8f1-kube-api-access-rhr9c\") pod \"nmstate-metrics-7f946cbc9-wgrqq\" (UID: \"90fcff6a-9f08-479c-8e5b-2745700bc8f1\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756381 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756444 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdftm\" (UniqueName: \"kubernetes.io/projected/8d2164f8-732d-4650-b92a-3fb677b51c25-kube-api-access-wdftm\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.756612 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/feed3d1f-a072-4ccc-b47a-3a5b48b44177-dbus-socket\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.773345 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhr9c\" (UniqueName: \"kubernetes.io/projected/90fcff6a-9f08-479c-8e5b-2745700bc8f1-kube-api-access-rhr9c\") pod \"nmstate-metrics-7f946cbc9-wgrqq\" (UID: \"90fcff6a-9f08-479c-8e5b-2745700bc8f1\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.782379 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npmks\" (UniqueName: \"kubernetes.io/projected/feed3d1f-a072-4ccc-b47a-3a5b48b44177-kube-api-access-npmks\") pod \"nmstate-handler-kgbh6\" (UID: \"feed3d1f-a072-4ccc-b47a-3a5b48b44177\") " pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.783138 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdftm\" (UniqueName: \"kubernetes.io/projected/8d2164f8-732d-4650-b92a-3fb677b51c25-kube-api-access-wdftm\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.857787 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.857897 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.858011 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc2rf\" (UniqueName: \"kubernetes.io/projected/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-kube-api-access-fc2rf\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.859734 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.864240 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.868215 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-69548fc9d4-657hq"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.868955 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.880305 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-69548fc9d4-657hq"] Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.880524 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc2rf\" (UniqueName: \"kubernetes.io/projected/f9dc9469-cf27-4d53-ae6d-a4df72d8af78-kube-api-access-fc2rf\") pod \"nmstate-console-plugin-7fbb5f6569-hzvx2\" (UID: \"f9dc9469-cf27-4d53-ae6d-a4df72d8af78\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.886898 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.928368 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:14 crc kubenswrapper[4817]: W1128 14:40:14.949495 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeed3d1f_a072_4ccc_b47a_3a5b48b44177.slice/crio-0615d8261c48ba2b34ade9f65d67955b87032125329dd99bdd75783bbcd2092c WatchSource:0}: Error finding container 0615d8261c48ba2b34ade9f65d67955b87032125329dd99bdd75783bbcd2092c: Status 404 returned error can't find the container with id 0615d8261c48ba2b34ade9f65d67955b87032125329dd99bdd75783bbcd2092c Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959616 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-service-ca\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959671 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-console-config\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959693 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq2d7\" (UniqueName: \"kubernetes.io/projected/7282c0dc-d0e7-4bdb-a378-319a34399055-kube-api-access-bq2d7\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959735 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-trusted-ca-bundle\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959765 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7282c0dc-d0e7-4bdb-a378-319a34399055-console-serving-cert\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959782 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7282c0dc-d0e7-4bdb-a378-319a34399055-console-oauth-config\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:14 crc kubenswrapper[4817]: I1128 14:40:14.959866 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-oauth-serving-cert\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.003858 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061492 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7282c0dc-d0e7-4bdb-a378-319a34399055-console-serving-cert\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061542 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7282c0dc-d0e7-4bdb-a378-319a34399055-console-oauth-config\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061572 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-oauth-serving-cert\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061646 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-service-ca\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061867 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-console-config\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061894 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq2d7\" (UniqueName: \"kubernetes.io/projected/7282c0dc-d0e7-4bdb-a378-319a34399055-kube-api-access-bq2d7\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.061928 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-trusted-ca-bundle\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.063005 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-service-ca\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.063081 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-console-config\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.063184 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-trusted-ca-bundle\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.063872 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7282c0dc-d0e7-4bdb-a378-319a34399055-oauth-serving-cert\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.070271 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7282c0dc-d0e7-4bdb-a378-319a34399055-console-oauth-config\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.070425 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7282c0dc-d0e7-4bdb-a378-319a34399055-console-serving-cert\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.082534 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq2d7\" (UniqueName: \"kubernetes.io/projected/7282c0dc-d0e7-4bdb-a378-319a34399055-kube-api-access-bq2d7\") pod \"console-69548fc9d4-657hq\" (UID: \"7282c0dc-d0e7-4bdb-a378-319a34399055\") " pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.244175 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.264844 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8d2164f8-732d-4650-b92a-3fb677b51c25-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.271993 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8d2164f8-732d-4650-b92a-3fb677b51c25-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-pm78n\" (UID: \"8d2164f8-732d-4650-b92a-3fb677b51c25\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.298577 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq"] Nov 28 14:40:15 crc kubenswrapper[4817]: W1128 14:40:15.304821 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90fcff6a_9f08_479c_8e5b_2745700bc8f1.slice/crio-1d4cb55f1a54f76e09adb7dac367b64f5818337c50794bcc91bf4f83bafb7c2b WatchSource:0}: Error finding container 1d4cb55f1a54f76e09adb7dac367b64f5818337c50794bcc91bf4f83bafb7c2b: Status 404 returned error can't find the container with id 1d4cb55f1a54f76e09adb7dac367b64f5818337c50794bcc91bf4f83bafb7c2b Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.408116 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2"] Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.449899 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-69548fc9d4-657hq"] Nov 28 14:40:15 crc kubenswrapper[4817]: W1128 14:40:15.453108 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7282c0dc_d0e7_4bdb_a378_319a34399055.slice/crio-5c031e7b4958dfb8f674274bea73216431a4bdb63a0cebd35d43863dca6009a8 WatchSource:0}: Error finding container 5c031e7b4958dfb8f674274bea73216431a4bdb63a0cebd35d43863dca6009a8: Status 404 returned error can't find the container with id 5c031e7b4958dfb8f674274bea73216431a4bdb63a0cebd35d43863dca6009a8 Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.499313 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.642071 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69548fc9d4-657hq" event={"ID":"7282c0dc-d0e7-4bdb-a378-319a34399055","Type":"ContainerStarted","Data":"8982720991d00a68b4afa09b62e161d653d3ad7d14ea8732ee7b31ee63f7bcb8"} Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.642164 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69548fc9d4-657hq" event={"ID":"7282c0dc-d0e7-4bdb-a378-319a34399055","Type":"ContainerStarted","Data":"5c031e7b4958dfb8f674274bea73216431a4bdb63a0cebd35d43863dca6009a8"} Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.643330 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kgbh6" event={"ID":"feed3d1f-a072-4ccc-b47a-3a5b48b44177","Type":"ContainerStarted","Data":"0615d8261c48ba2b34ade9f65d67955b87032125329dd99bdd75783bbcd2092c"} Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.644416 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" event={"ID":"f9dc9469-cf27-4d53-ae6d-a4df72d8af78","Type":"ContainerStarted","Data":"7678a96dd5fd4cd404c93f1ab8cb7ce64e41edbd65692af928e3009c13c2f4b1"} Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.645227 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" event={"ID":"90fcff6a-9f08-479c-8e5b-2745700bc8f1","Type":"ContainerStarted","Data":"1d4cb55f1a54f76e09adb7dac367b64f5818337c50794bcc91bf4f83bafb7c2b"} Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.667766 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-69548fc9d4-657hq" podStartSLOduration=1.6677443429999999 podStartE2EDuration="1.667744343s" podCreationTimestamp="2025-11-28 14:40:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:40:15.665394455 +0000 UTC m=+638.253372721" watchObservedRunningTime="2025-11-28 14:40:15.667744343 +0000 UTC m=+638.255722609" Nov 28 14:40:15 crc kubenswrapper[4817]: I1128 14:40:15.721137 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n"] Nov 28 14:40:16 crc kubenswrapper[4817]: I1128 14:40:16.652063 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" event={"ID":"8d2164f8-732d-4650-b92a-3fb677b51c25","Type":"ContainerStarted","Data":"e981f562ba0bf0ab071d9b65120be9861aa4385942d289cb1e2b04a54dca8ee2"} Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.665832 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kgbh6" event={"ID":"feed3d1f-a072-4ccc-b47a-3a5b48b44177","Type":"ContainerStarted","Data":"b5185c6c47256bbaa6ef85711e5ac475b4ba3ea1e1a1fc13bdaa908ff05b2d1c"} Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.666163 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.668583 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" event={"ID":"8d2164f8-732d-4650-b92a-3fb677b51c25","Type":"ContainerStarted","Data":"db8b744ef046dcec5dacdf114f172704408eb3803dbf334d8a0900dd230c89de"} Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.668694 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.673774 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" event={"ID":"f9dc9469-cf27-4d53-ae6d-a4df72d8af78","Type":"ContainerStarted","Data":"d63c46ab8562e05dde6245d1e60d23a80579b4c7889a7b5b6032d16fd26a9766"} Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.684274 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kgbh6" podStartSLOduration=1.650032986 podStartE2EDuration="4.684254862s" podCreationTimestamp="2025-11-28 14:40:14 +0000 UTC" firstStartedPulling="2025-11-28 14:40:14.952209915 +0000 UTC m=+637.540188181" lastFinishedPulling="2025-11-28 14:40:17.986431801 +0000 UTC m=+640.574410057" observedRunningTime="2025-11-28 14:40:18.683415931 +0000 UTC m=+641.271394207" watchObservedRunningTime="2025-11-28 14:40:18.684254862 +0000 UTC m=+641.272233128" Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.701797 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" podStartSLOduration=2.443210279 podStartE2EDuration="4.701780654s" podCreationTimestamp="2025-11-28 14:40:14 +0000 UTC" firstStartedPulling="2025-11-28 14:40:15.728919182 +0000 UTC m=+638.316897448" lastFinishedPulling="2025-11-28 14:40:17.987489557 +0000 UTC m=+640.575467823" observedRunningTime="2025-11-28 14:40:18.697426377 +0000 UTC m=+641.285404653" watchObservedRunningTime="2025-11-28 14:40:18.701780654 +0000 UTC m=+641.289758920" Nov 28 14:40:18 crc kubenswrapper[4817]: I1128 14:40:18.717020 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hzvx2" podStartSLOduration=2.153711709 podStartE2EDuration="4.71700364s" podCreationTimestamp="2025-11-28 14:40:14 +0000 UTC" firstStartedPulling="2025-11-28 14:40:15.419517461 +0000 UTC m=+638.007495737" lastFinishedPulling="2025-11-28 14:40:17.982809372 +0000 UTC m=+640.570787668" observedRunningTime="2025-11-28 14:40:18.712595281 +0000 UTC m=+641.300573547" watchObservedRunningTime="2025-11-28 14:40:18.71700364 +0000 UTC m=+641.304981906" Nov 28 14:40:19 crc kubenswrapper[4817]: I1128 14:40:19.682655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" event={"ID":"90fcff6a-9f08-479c-8e5b-2745700bc8f1","Type":"ContainerStarted","Data":"26571fddc3251cd7371be67dda99431391cb5330d390e1ace10fc4603db69325"} Nov 28 14:40:21 crc kubenswrapper[4817]: I1128 14:40:21.697929 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" event={"ID":"90fcff6a-9f08-479c-8e5b-2745700bc8f1","Type":"ContainerStarted","Data":"abe34302612cd83f113f497bed26ce1a95d126fabdfe62d41fb5845125de4dc9"} Nov 28 14:40:21 crc kubenswrapper[4817]: I1128 14:40:21.717930 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wgrqq" podStartSLOduration=2.006897078 podStartE2EDuration="7.717911254s" podCreationTimestamp="2025-11-28 14:40:14 +0000 UTC" firstStartedPulling="2025-11-28 14:40:15.308431901 +0000 UTC m=+637.896410167" lastFinishedPulling="2025-11-28 14:40:21.019446077 +0000 UTC m=+643.607424343" observedRunningTime="2025-11-28 14:40:21.714481359 +0000 UTC m=+644.302459625" watchObservedRunningTime="2025-11-28 14:40:21.717911254 +0000 UTC m=+644.305889510" Nov 28 14:40:24 crc kubenswrapper[4817]: I1128 14:40:24.961428 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kgbh6" Nov 28 14:40:25 crc kubenswrapper[4817]: I1128 14:40:25.245236 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:25 crc kubenswrapper[4817]: I1128 14:40:25.245309 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:25 crc kubenswrapper[4817]: I1128 14:40:25.249779 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:25 crc kubenswrapper[4817]: I1128 14:40:25.729694 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-69548fc9d4-657hq" Nov 28 14:40:25 crc kubenswrapper[4817]: I1128 14:40:25.797443 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9xjrk"] Nov 28 14:40:35 crc kubenswrapper[4817]: I1128 14:40:35.508500 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-pm78n" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.329089 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5"] Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.331304 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.336910 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.347952 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5"] Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.465322 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.465384 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spwfm\" (UniqueName: \"kubernetes.io/projected/645c8f43-6461-4c60-8fca-ccdde864e321-kube-api-access-spwfm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.465408 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.567688 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.567991 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spwfm\" (UniqueName: \"kubernetes.io/projected/645c8f43-6461-4c60-8fca-ccdde864e321-kube-api-access-spwfm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.568021 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.568602 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.568651 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.593857 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spwfm\" (UniqueName: \"kubernetes.io/projected/645c8f43-6461-4c60-8fca-ccdde864e321-kube-api-access-spwfm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:49 crc kubenswrapper[4817]: I1128 14:40:49.664815 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:50 crc kubenswrapper[4817]: I1128 14:40:50.091014 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5"] Nov 28 14:40:50 crc kubenswrapper[4817]: W1128 14:40:50.097123 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod645c8f43_6461_4c60_8fca_ccdde864e321.slice/crio-0e81b605ae648e6f7195845a097c81cabaa6095befe827f84a7006a2c57a5e28 WatchSource:0}: Error finding container 0e81b605ae648e6f7195845a097c81cabaa6095befe827f84a7006a2c57a5e28: Status 404 returned error can't find the container with id 0e81b605ae648e6f7195845a097c81cabaa6095befe827f84a7006a2c57a5e28 Nov 28 14:40:50 crc kubenswrapper[4817]: I1128 14:40:50.866444 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-9xjrk" podUID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" containerName="console" containerID="cri-o://b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae" gracePeriod=15 Nov 28 14:40:50 crc kubenswrapper[4817]: I1128 14:40:50.896629 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" event={"ID":"645c8f43-6461-4c60-8fca-ccdde864e321","Type":"ContainerStarted","Data":"0e81b605ae648e6f7195845a097c81cabaa6095befe827f84a7006a2c57a5e28"} Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.203177 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9xjrk_8d7e9659-3a05-461d-ab42-ceda65eb40e9/console/0.log" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.203429 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.391867 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-service-ca\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.391993 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-oauth-serving-cert\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.392030 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krwn6\" (UniqueName: \"kubernetes.io/projected/8d7e9659-3a05-461d-ab42-ceda65eb40e9-kube-api-access-krwn6\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.392067 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-config\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.392106 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-trusted-ca-bundle\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.392155 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-oauth-config\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.392198 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-serving-cert\") pod \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\" (UID: \"8d7e9659-3a05-461d-ab42-ceda65eb40e9\") " Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.393200 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.393211 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.393269 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-config" (OuterVolumeSpecName: "console-config") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.393682 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-service-ca" (OuterVolumeSpecName: "service-ca") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.399906 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.400205 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.401029 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7e9659-3a05-461d-ab42-ceda65eb40e9-kube-api-access-krwn6" (OuterVolumeSpecName: "kube-api-access-krwn6") pod "8d7e9659-3a05-461d-ab42-ceda65eb40e9" (UID: "8d7e9659-3a05-461d-ab42-ceda65eb40e9"). InnerVolumeSpecName "kube-api-access-krwn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494220 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krwn6\" (UniqueName: \"kubernetes.io/projected/8d7e9659-3a05-461d-ab42-ceda65eb40e9-kube-api-access-krwn6\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494263 4817 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494275 4817 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494286 4817 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494298 4817 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d7e9659-3a05-461d-ab42-ceda65eb40e9-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494309 4817 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.494322 4817 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d7e9659-3a05-461d-ab42-ceda65eb40e9-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.909315 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9xjrk_8d7e9659-3a05-461d-ab42-ceda65eb40e9/console/0.log" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.909396 4817 generic.go:334] "Generic (PLEG): container finished" podID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" containerID="b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae" exitCode=2 Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.909485 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9xjrk" event={"ID":"8d7e9659-3a05-461d-ab42-ceda65eb40e9","Type":"ContainerDied","Data":"b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae"} Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.909523 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9xjrk" event={"ID":"8d7e9659-3a05-461d-ab42-ceda65eb40e9","Type":"ContainerDied","Data":"6b066455e48ca3fe45114ae75afec09264630928cbdd64ed5f474df82d340094"} Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.909541 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9xjrk" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.909551 4817 scope.go:117] "RemoveContainer" containerID="b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.913968 4817 generic.go:334] "Generic (PLEG): container finished" podID="645c8f43-6461-4c60-8fca-ccdde864e321" containerID="2e52e4a5aadf64d5e90ab3b7a6eba61835f4e964fe2bf0fe592e99f4882e92df" exitCode=0 Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.914023 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" event={"ID":"645c8f43-6461-4c60-8fca-ccdde864e321","Type":"ContainerDied","Data":"2e52e4a5aadf64d5e90ab3b7a6eba61835f4e964fe2bf0fe592e99f4882e92df"} Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.944003 4817 scope.go:117] "RemoveContainer" containerID="b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.946769 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9xjrk"] Nov 28 14:40:51 crc kubenswrapper[4817]: E1128 14:40:51.951315 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae\": container with ID starting with b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae not found: ID does not exist" containerID="b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.951375 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae"} err="failed to get container status \"b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae\": rpc error: code = NotFound desc = could not find container \"b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae\": container with ID starting with b3fba65b2896f1920ce16d90f0bd5de2fc843a061a8fb1e1f7af300e9fff50ae not found: ID does not exist" Nov 28 14:40:51 crc kubenswrapper[4817]: I1128 14:40:51.956421 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-9xjrk"] Nov 28 14:40:53 crc kubenswrapper[4817]: I1128 14:40:53.787774 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" path="/var/lib/kubelet/pods/8d7e9659-3a05-461d-ab42-ceda65eb40e9/volumes" Nov 28 14:40:53 crc kubenswrapper[4817]: I1128 14:40:53.928961 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" event={"ID":"645c8f43-6461-4c60-8fca-ccdde864e321","Type":"ContainerStarted","Data":"d8d296484de551069a1937fb89d2799bb88a44088d57189bcc57c07ed9f2cbe3"} Nov 28 14:40:54 crc kubenswrapper[4817]: I1128 14:40:54.936071 4817 generic.go:334] "Generic (PLEG): container finished" podID="645c8f43-6461-4c60-8fca-ccdde864e321" containerID="d8d296484de551069a1937fb89d2799bb88a44088d57189bcc57c07ed9f2cbe3" exitCode=0 Nov 28 14:40:54 crc kubenswrapper[4817]: I1128 14:40:54.936140 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" event={"ID":"645c8f43-6461-4c60-8fca-ccdde864e321","Type":"ContainerDied","Data":"d8d296484de551069a1937fb89d2799bb88a44088d57189bcc57c07ed9f2cbe3"} Nov 28 14:40:55 crc kubenswrapper[4817]: I1128 14:40:55.944252 4817 generic.go:334] "Generic (PLEG): container finished" podID="645c8f43-6461-4c60-8fca-ccdde864e321" containerID="efad6fdedb77296710bd482957e77f2b016f0f0ac8246c3322f17bb4eaeebcc8" exitCode=0 Nov 28 14:40:55 crc kubenswrapper[4817]: I1128 14:40:55.944325 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" event={"ID":"645c8f43-6461-4c60-8fca-ccdde864e321","Type":"ContainerDied","Data":"efad6fdedb77296710bd482957e77f2b016f0f0ac8246c3322f17bb4eaeebcc8"} Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.178870 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.369298 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-util\") pod \"645c8f43-6461-4c60-8fca-ccdde864e321\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.369408 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-bundle\") pod \"645c8f43-6461-4c60-8fca-ccdde864e321\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.369501 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spwfm\" (UniqueName: \"kubernetes.io/projected/645c8f43-6461-4c60-8fca-ccdde864e321-kube-api-access-spwfm\") pod \"645c8f43-6461-4c60-8fca-ccdde864e321\" (UID: \"645c8f43-6461-4c60-8fca-ccdde864e321\") " Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.370611 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-bundle" (OuterVolumeSpecName: "bundle") pod "645c8f43-6461-4c60-8fca-ccdde864e321" (UID: "645c8f43-6461-4c60-8fca-ccdde864e321"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.379315 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645c8f43-6461-4c60-8fca-ccdde864e321-kube-api-access-spwfm" (OuterVolumeSpecName: "kube-api-access-spwfm") pod "645c8f43-6461-4c60-8fca-ccdde864e321" (UID: "645c8f43-6461-4c60-8fca-ccdde864e321"). InnerVolumeSpecName "kube-api-access-spwfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.397852 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-util" (OuterVolumeSpecName: "util") pod "645c8f43-6461-4c60-8fca-ccdde864e321" (UID: "645c8f43-6461-4c60-8fca-ccdde864e321"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.471225 4817 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-util\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.471294 4817 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/645c8f43-6461-4c60-8fca-ccdde864e321-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.471332 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spwfm\" (UniqueName: \"kubernetes.io/projected/645c8f43-6461-4c60-8fca-ccdde864e321-kube-api-access-spwfm\") on node \"crc\" DevicePath \"\"" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.956594 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" event={"ID":"645c8f43-6461-4c60-8fca-ccdde864e321","Type":"ContainerDied","Data":"0e81b605ae648e6f7195845a097c81cabaa6095befe827f84a7006a2c57a5e28"} Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.956647 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e81b605ae648e6f7195845a097c81cabaa6095befe827f84a7006a2c57a5e28" Nov 28 14:40:57 crc kubenswrapper[4817]: I1128 14:40:57.956704 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.929506 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj"] Nov 28 14:41:07 crc kubenswrapper[4817]: E1128 14:41:07.930315 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" containerName="console" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.930331 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" containerName="console" Nov 28 14:41:07 crc kubenswrapper[4817]: E1128 14:41:07.930352 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="util" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.930359 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="util" Nov 28 14:41:07 crc kubenswrapper[4817]: E1128 14:41:07.930369 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="pull" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.930375 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="pull" Nov 28 14:41:07 crc kubenswrapper[4817]: E1128 14:41:07.930385 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="extract" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.930391 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="extract" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.930514 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7e9659-3a05-461d-ab42-ceda65eb40e9" containerName="console" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.930533 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="645c8f43-6461-4c60-8fca-ccdde864e321" containerName="extract" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.931003 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.933165 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.933634 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9b9rl" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.933670 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.934340 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.938916 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 28 14:41:07 crc kubenswrapper[4817]: I1128 14:41:07.990565 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj"] Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.009252 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aebf367a-7ec1-4675-b01f-3d54d2ba0613-webhook-cert\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.009301 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26vj5\" (UniqueName: \"kubernetes.io/projected/aebf367a-7ec1-4675-b01f-3d54d2ba0613-kube-api-access-26vj5\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.009442 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aebf367a-7ec1-4675-b01f-3d54d2ba0613-apiservice-cert\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.110253 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aebf367a-7ec1-4675-b01f-3d54d2ba0613-apiservice-cert\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.110314 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aebf367a-7ec1-4675-b01f-3d54d2ba0613-webhook-cert\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.110351 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26vj5\" (UniqueName: \"kubernetes.io/projected/aebf367a-7ec1-4675-b01f-3d54d2ba0613-kube-api-access-26vj5\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.118172 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aebf367a-7ec1-4675-b01f-3d54d2ba0613-apiservice-cert\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.119234 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aebf367a-7ec1-4675-b01f-3d54d2ba0613-webhook-cert\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.140139 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26vj5\" (UniqueName: \"kubernetes.io/projected/aebf367a-7ec1-4675-b01f-3d54d2ba0613-kube-api-access-26vj5\") pod \"metallb-operator-controller-manager-5d674c8684-kk2kj\" (UID: \"aebf367a-7ec1-4675-b01f-3d54d2ba0613\") " pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.180533 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4"] Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.181393 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.184694 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.185004 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jvsp2" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.185143 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.202121 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4"] Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.212479 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-webhook-cert\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.212568 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-apiservice-cert\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.212622 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jql9k\" (UniqueName: \"kubernetes.io/projected/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-kube-api-access-jql9k\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.249636 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.313498 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jql9k\" (UniqueName: \"kubernetes.io/projected/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-kube-api-access-jql9k\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.313854 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-webhook-cert\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.313915 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-apiservice-cert\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.332118 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-webhook-cert\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.332121 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-apiservice-cert\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.338144 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jql9k\" (UniqueName: \"kubernetes.io/projected/5980acb2-a5d3-44fa-8bf0-23e36cde74a4-kube-api-access-jql9k\") pod \"metallb-operator-webhook-server-74f997fcc-g6nm4\" (UID: \"5980acb2-a5d3-44fa-8bf0-23e36cde74a4\") " pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.504799 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:08 crc kubenswrapper[4817]: I1128 14:41:08.529018 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj"] Nov 28 14:41:08 crc kubenswrapper[4817]: W1128 14:41:08.546860 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaebf367a_7ec1_4675_b01f_3d54d2ba0613.slice/crio-4b68faed0e9a79e4062ddf6276e737a1e1c6e9c6f537ec4be36fa9632b7c9ff1 WatchSource:0}: Error finding container 4b68faed0e9a79e4062ddf6276e737a1e1c6e9c6f537ec4be36fa9632b7c9ff1: Status 404 returned error can't find the container with id 4b68faed0e9a79e4062ddf6276e737a1e1c6e9c6f537ec4be36fa9632b7c9ff1 Nov 28 14:41:09 crc kubenswrapper[4817]: I1128 14:41:09.015710 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4"] Nov 28 14:41:09 crc kubenswrapper[4817]: I1128 14:41:09.018134 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" event={"ID":"aebf367a-7ec1-4675-b01f-3d54d2ba0613","Type":"ContainerStarted","Data":"4b68faed0e9a79e4062ddf6276e737a1e1c6e9c6f537ec4be36fa9632b7c9ff1"} Nov 28 14:41:10 crc kubenswrapper[4817]: I1128 14:41:10.034280 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" event={"ID":"5980acb2-a5d3-44fa-8bf0-23e36cde74a4","Type":"ContainerStarted","Data":"024563e300fba6adec0849ea4a5cd9b30d808cf74ba6490b3d1f74937ca1205e"} Nov 28 14:41:12 crc kubenswrapper[4817]: I1128 14:41:12.055986 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" event={"ID":"aebf367a-7ec1-4675-b01f-3d54d2ba0613","Type":"ContainerStarted","Data":"d95ac5ab9dbebd256a03312668fa95d6db4f388ca0f65400731dc980f5c701cb"} Nov 28 14:41:12 crc kubenswrapper[4817]: I1128 14:41:12.056415 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:12 crc kubenswrapper[4817]: I1128 14:41:12.082000 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" podStartSLOduration=2.11282237 podStartE2EDuration="5.081979982s" podCreationTimestamp="2025-11-28 14:41:07 +0000 UTC" firstStartedPulling="2025-11-28 14:41:08.54881708 +0000 UTC m=+691.136795346" lastFinishedPulling="2025-11-28 14:41:11.517974692 +0000 UTC m=+694.105952958" observedRunningTime="2025-11-28 14:41:12.07822954 +0000 UTC m=+694.666207806" watchObservedRunningTime="2025-11-28 14:41:12.081979982 +0000 UTC m=+694.669958238" Nov 28 14:41:14 crc kubenswrapper[4817]: I1128 14:41:14.068225 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" event={"ID":"5980acb2-a5d3-44fa-8bf0-23e36cde74a4","Type":"ContainerStarted","Data":"e4df116022a3000398ae1ce511568912941263867b99b37cd62ba280eda584e8"} Nov 28 14:41:14 crc kubenswrapper[4817]: I1128 14:41:14.068591 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:14 crc kubenswrapper[4817]: I1128 14:41:14.103242 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" podStartSLOduration=1.926707615 podStartE2EDuration="6.103219925s" podCreationTimestamp="2025-11-28 14:41:08 +0000 UTC" firstStartedPulling="2025-11-28 14:41:09.026708947 +0000 UTC m=+691.614687213" lastFinishedPulling="2025-11-28 14:41:13.203221237 +0000 UTC m=+695.791199523" observedRunningTime="2025-11-28 14:41:14.101162784 +0000 UTC m=+696.689141050" watchObservedRunningTime="2025-11-28 14:41:14.103219925 +0000 UTC m=+696.691198191" Nov 28 14:41:28 crc kubenswrapper[4817]: I1128 14:41:28.513180 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-74f997fcc-g6nm4" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.253780 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d674c8684-kk2kj" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.951858 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mfq4b"] Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.954036 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.957573 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.957622 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.957858 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-d7zfc" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.975047 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x"] Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.975857 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.978485 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 28 14:41:48 crc kubenswrapper[4817]: I1128 14:41:48.986333 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x"] Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.043999 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-metrics\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044040 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6mq7\" (UniqueName: \"kubernetes.io/projected/39068359-e185-4366-9378-718043d84ba2-kube-api-access-c6mq7\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044136 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-reloader\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044184 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39068359-e185-4366-9378-718043d84ba2-metrics-certs\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044252 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-frr-conf\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044287 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/39068359-e185-4366-9378-718043d84ba2-frr-startup\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044314 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqjwl\" (UniqueName: \"kubernetes.io/projected/8f400971-fd8a-4c48-82bd-33f47438bcb0-kube-api-access-xqjwl\") pod \"frr-k8s-webhook-server-7fcb986d4-4r56x\" (UID: \"8f400971-fd8a-4c48-82bd-33f47438bcb0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044369 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f400971-fd8a-4c48-82bd-33f47438bcb0-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-4r56x\" (UID: \"8f400971-fd8a-4c48-82bd-33f47438bcb0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.044401 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-frr-sockets\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.060682 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hkbsn"] Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.061802 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.063947 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.064049 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-59g7c" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.064150 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.064432 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.074905 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-4wsbn"] Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.075935 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.077633 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.090120 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-4wsbn"] Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145103 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39068359-e185-4366-9378-718043d84ba2-metrics-certs\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145151 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-frr-conf\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145167 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/39068359-e185-4366-9378-718043d84ba2-frr-startup\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145187 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqjwl\" (UniqueName: \"kubernetes.io/projected/8f400971-fd8a-4c48-82bd-33f47438bcb0-kube-api-access-xqjwl\") pod \"frr-k8s-webhook-server-7fcb986d4-4r56x\" (UID: \"8f400971-fd8a-4c48-82bd-33f47438bcb0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145210 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metrics-certs\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145232 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metallb-excludel2\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145246 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m28nh\" (UniqueName: \"kubernetes.io/projected/f737005e-fcc0-4f24-bdbb-99d7f22afef6-kube-api-access-m28nh\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145296 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f400971-fd8a-4c48-82bd-33f47438bcb0-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-4r56x\" (UID: \"8f400971-fd8a-4c48-82bd-33f47438bcb0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145317 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-frr-sockets\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145365 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-metrics-certs\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145386 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-cert\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145407 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-metrics\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.145433 4817 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145450 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6mq7\" (UniqueName: \"kubernetes.io/projected/39068359-e185-4366-9378-718043d84ba2-kube-api-access-c6mq7\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.145527 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39068359-e185-4366-9378-718043d84ba2-metrics-certs podName:39068359-e185-4366-9378-718043d84ba2 nodeName:}" failed. No retries permitted until 2025-11-28 14:41:49.645505406 +0000 UTC m=+732.233483682 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/39068359-e185-4366-9378-718043d84ba2-metrics-certs") pod "frr-k8s-mfq4b" (UID: "39068359-e185-4366-9378-718043d84ba2") : secret "frr-k8s-certs-secret" not found Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145567 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s6s7\" (UniqueName: \"kubernetes.io/projected/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-kube-api-access-2s6s7\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145578 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-frr-conf\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145605 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145641 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-reloader\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145707 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-metrics\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145900 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-reloader\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.145935 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/39068359-e185-4366-9378-718043d84ba2-frr-sockets\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.146353 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/39068359-e185-4366-9378-718043d84ba2-frr-startup\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.162445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f400971-fd8a-4c48-82bd-33f47438bcb0-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-4r56x\" (UID: \"8f400971-fd8a-4c48-82bd-33f47438bcb0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.171484 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6mq7\" (UniqueName: \"kubernetes.io/projected/39068359-e185-4366-9378-718043d84ba2-kube-api-access-c6mq7\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.171670 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqjwl\" (UniqueName: \"kubernetes.io/projected/8f400971-fd8a-4c48-82bd-33f47438bcb0-kube-api-access-xqjwl\") pod \"frr-k8s-webhook-server-7fcb986d4-4r56x\" (UID: \"8f400971-fd8a-4c48-82bd-33f47438bcb0\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246510 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metrics-certs\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246581 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metallb-excludel2\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246602 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m28nh\" (UniqueName: \"kubernetes.io/projected/f737005e-fcc0-4f24-bdbb-99d7f22afef6-kube-api-access-m28nh\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246636 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-metrics-certs\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246661 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-cert\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.246687 4817 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.246790 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metrics-certs podName:f737005e-fcc0-4f24-bdbb-99d7f22afef6 nodeName:}" failed. No retries permitted until 2025-11-28 14:41:49.746766334 +0000 UTC m=+732.334744600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metrics-certs") pod "speaker-hkbsn" (UID: "f737005e-fcc0-4f24-bdbb-99d7f22afef6") : secret "speaker-certs-secret" not found Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246701 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s6s7\" (UniqueName: \"kubernetes.io/projected/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-kube-api-access-2s6s7\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.246964 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.247104 4817 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.247145 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-metrics-certs podName:7f4b2cee-aa3d-43ca-adbd-42dee238c7bc nodeName:}" failed. No retries permitted until 2025-11-28 14:41:49.747134893 +0000 UTC m=+732.335113289 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-metrics-certs") pod "controller-f8648f98b-4wsbn" (UID: "7f4b2cee-aa3d-43ca-adbd-42dee238c7bc") : secret "controller-certs-secret" not found Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.247209 4817 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.247280 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist podName:f737005e-fcc0-4f24-bdbb-99d7f22afef6 nodeName:}" failed. No retries permitted until 2025-11-28 14:41:49.747261336 +0000 UTC m=+732.335239702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist") pod "speaker-hkbsn" (UID: "f737005e-fcc0-4f24-bdbb-99d7f22afef6") : secret "metallb-memberlist" not found Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.247504 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metallb-excludel2\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.248840 4817 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.263479 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-cert\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.266004 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s6s7\" (UniqueName: \"kubernetes.io/projected/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-kube-api-access-2s6s7\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.269182 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m28nh\" (UniqueName: \"kubernetes.io/projected/f737005e-fcc0-4f24-bdbb-99d7f22afef6-kube-api-access-m28nh\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.289259 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.515419 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x"] Nov 28 14:41:49 crc kubenswrapper[4817]: W1128 14:41:49.518848 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f400971_fd8a_4c48_82bd_33f47438bcb0.slice/crio-99b6aed690d4aa0afdbc59c0d44e68df76bf7e22398e9aafde1e3f38c83c189f WatchSource:0}: Error finding container 99b6aed690d4aa0afdbc59c0d44e68df76bf7e22398e9aafde1e3f38c83c189f: Status 404 returned error can't find the container with id 99b6aed690d4aa0afdbc59c0d44e68df76bf7e22398e9aafde1e3f38c83c189f Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.650898 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39068359-e185-4366-9378-718043d84ba2-metrics-certs\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.655629 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/39068359-e185-4366-9378-718043d84ba2-metrics-certs\") pod \"frr-k8s-mfq4b\" (UID: \"39068359-e185-4366-9378-718043d84ba2\") " pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.754435 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.754555 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metrics-certs\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.754644 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-metrics-certs\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.755851 4817 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 14:41:49 crc kubenswrapper[4817]: E1128 14:41:49.755948 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist podName:f737005e-fcc0-4f24-bdbb-99d7f22afef6 nodeName:}" failed. No retries permitted until 2025-11-28 14:41:50.755924572 +0000 UTC m=+733.343902838 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist") pod "speaker-hkbsn" (UID: "f737005e-fcc0-4f24-bdbb-99d7f22afef6") : secret "metallb-memberlist" not found Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.771925 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f4b2cee-aa3d-43ca-adbd-42dee238c7bc-metrics-certs\") pod \"controller-f8648f98b-4wsbn\" (UID: \"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc\") " pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.776042 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-metrics-certs\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.869477 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:41:49 crc kubenswrapper[4817]: I1128 14:41:49.989904 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.171939 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-4wsbn"] Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.288204 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" event={"ID":"8f400971-fd8a-4c48-82bd-33f47438bcb0","Type":"ContainerStarted","Data":"99b6aed690d4aa0afdbc59c0d44e68df76bf7e22398e9aafde1e3f38c83c189f"} Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.288810 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"7cd1d3bafb9f50450e1273649162093b3de94af8a93db2a672dfd9420855e79d"} Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.289742 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-4wsbn" event={"ID":"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc","Type":"ContainerStarted","Data":"009f87310527bc2bc42c2404b95a59e6915923ce6933f53f7e2cd39344b203fa"} Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.769020 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.776022 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f737005e-fcc0-4f24-bdbb-99d7f22afef6-memberlist\") pod \"speaker-hkbsn\" (UID: \"f737005e-fcc0-4f24-bdbb-99d7f22afef6\") " pod="metallb-system/speaker-hkbsn" Nov 28 14:41:50 crc kubenswrapper[4817]: I1128 14:41:50.876340 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hkbsn" Nov 28 14:41:50 crc kubenswrapper[4817]: W1128 14:41:50.902860 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf737005e_fcc0_4f24_bdbb_99d7f22afef6.slice/crio-1a8a0e4e196456a280d16ae612e23584670e0382fe547bf02ba331487c523e42 WatchSource:0}: Error finding container 1a8a0e4e196456a280d16ae612e23584670e0382fe547bf02ba331487c523e42: Status 404 returned error can't find the container with id 1a8a0e4e196456a280d16ae612e23584670e0382fe547bf02ba331487c523e42 Nov 28 14:41:51 crc kubenswrapper[4817]: I1128 14:41:51.305311 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hkbsn" event={"ID":"f737005e-fcc0-4f24-bdbb-99d7f22afef6","Type":"ContainerStarted","Data":"8f0f274bbf7a769d61a78db076eac48ebd4d6e22459217360938386f9015b431"} Nov 28 14:41:51 crc kubenswrapper[4817]: I1128 14:41:51.305363 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hkbsn" event={"ID":"f737005e-fcc0-4f24-bdbb-99d7f22afef6","Type":"ContainerStarted","Data":"1a8a0e4e196456a280d16ae612e23584670e0382fe547bf02ba331487c523e42"} Nov 28 14:41:51 crc kubenswrapper[4817]: I1128 14:41:51.307703 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-4wsbn" event={"ID":"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc","Type":"ContainerStarted","Data":"f3f231a643b41c7404ba38280f5193be73c9c0a60dab6c2b456b3a60d098d0c9"} Nov 28 14:41:51 crc kubenswrapper[4817]: I1128 14:41:51.307767 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-4wsbn" event={"ID":"7f4b2cee-aa3d-43ca-adbd-42dee238c7bc","Type":"ContainerStarted","Data":"acb93774aeccf521d3f40a979e48cdf2b5c9dde21337161de2537baeb4b28dc9"} Nov 28 14:41:51 crc kubenswrapper[4817]: I1128 14:41:51.308666 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:41:51 crc kubenswrapper[4817]: I1128 14:41:51.330367 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-4wsbn" podStartSLOduration=2.330321252 podStartE2EDuration="2.330321252s" podCreationTimestamp="2025-11-28 14:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:41:51.327264847 +0000 UTC m=+733.915243113" watchObservedRunningTime="2025-11-28 14:41:51.330321252 +0000 UTC m=+733.918299518" Nov 28 14:41:52 crc kubenswrapper[4817]: I1128 14:41:52.342208 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hkbsn" event={"ID":"f737005e-fcc0-4f24-bdbb-99d7f22afef6","Type":"ContainerStarted","Data":"aed44602491804dc9e8489a482b64de2bf41e36ceae76d26c0365d68ea32bf1a"} Nov 28 14:41:52 crc kubenswrapper[4817]: I1128 14:41:52.342605 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hkbsn" Nov 28 14:41:52 crc kubenswrapper[4817]: I1128 14:41:52.364356 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hkbsn" podStartSLOduration=3.3643321950000002 podStartE2EDuration="3.364332195s" podCreationTimestamp="2025-11-28 14:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:41:52.359065776 +0000 UTC m=+734.947044052" watchObservedRunningTime="2025-11-28 14:41:52.364332195 +0000 UTC m=+734.952310461" Nov 28 14:41:57 crc kubenswrapper[4817]: I1128 14:41:57.378097 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" event={"ID":"8f400971-fd8a-4c48-82bd-33f47438bcb0","Type":"ContainerStarted","Data":"b6c4e49a88899169266b369d9170c60f118db1eb0e043379bba0082a6bbc0e10"} Nov 28 14:41:57 crc kubenswrapper[4817]: I1128 14:41:57.379212 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:41:57 crc kubenswrapper[4817]: I1128 14:41:57.380873 4817 generic.go:334] "Generic (PLEG): container finished" podID="39068359-e185-4366-9378-718043d84ba2" containerID="45752990ce4acfcf7ae4670cc00248b95069d6433f2f03d1f5deb88dbeec2702" exitCode=0 Nov 28 14:41:57 crc kubenswrapper[4817]: I1128 14:41:57.380918 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerDied","Data":"45752990ce4acfcf7ae4670cc00248b95069d6433f2f03d1f5deb88dbeec2702"} Nov 28 14:41:57 crc kubenswrapper[4817]: I1128 14:41:57.400352 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" podStartSLOduration=2.156420709 podStartE2EDuration="9.400323903s" podCreationTimestamp="2025-11-28 14:41:48 +0000 UTC" firstStartedPulling="2025-11-28 14:41:49.520750631 +0000 UTC m=+732.108728917" lastFinishedPulling="2025-11-28 14:41:56.764653845 +0000 UTC m=+739.352632111" observedRunningTime="2025-11-28 14:41:57.393857433 +0000 UTC m=+739.981835699" watchObservedRunningTime="2025-11-28 14:41:57.400323903 +0000 UTC m=+739.988302189" Nov 28 14:41:58 crc kubenswrapper[4817]: I1128 14:41:58.389681 4817 generic.go:334] "Generic (PLEG): container finished" podID="39068359-e185-4366-9378-718043d84ba2" containerID="b790ef7a1b1ae138f88fe4c5556fdaf05054951823f136abc6e0578c793fff30" exitCode=0 Nov 28 14:41:58 crc kubenswrapper[4817]: I1128 14:41:58.389755 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerDied","Data":"b790ef7a1b1ae138f88fe4c5556fdaf05054951823f136abc6e0578c793fff30"} Nov 28 14:41:59 crc kubenswrapper[4817]: I1128 14:41:59.401200 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerDied","Data":"51cd79379a25a066c37823fd874211019d92601efdd8d671bef7719a54fc06c0"} Nov 28 14:41:59 crc kubenswrapper[4817]: I1128 14:41:59.401039 4817 generic.go:334] "Generic (PLEG): container finished" podID="39068359-e185-4366-9378-718043d84ba2" containerID="51cd79379a25a066c37823fd874211019d92601efdd8d671bef7719a54fc06c0" exitCode=0 Nov 28 14:42:00 crc kubenswrapper[4817]: I1128 14:42:00.410638 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"511e4be8200404bd15152ecb809479a264d75e8c01aef916604be27862de2f3f"} Nov 28 14:42:00 crc kubenswrapper[4817]: I1128 14:42:00.410988 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"f5cbead675166263b2e86f944a5c8e8a92e2aa7e9a1517152d8fb29d3ed82acf"} Nov 28 14:42:00 crc kubenswrapper[4817]: I1128 14:42:00.411004 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"ef1421de47ebc5d9c2115484d22f868316fdf6db8d1c0c9b7cf62a858ea9546b"} Nov 28 14:42:00 crc kubenswrapper[4817]: I1128 14:42:00.411014 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"5d5c3e52b90920759fb4bdb4308b71cbb5e2c4ad5bd914f56af90659756ee0a5"} Nov 28 14:42:00 crc kubenswrapper[4817]: I1128 14:42:00.411025 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"efb64d20eb87be92a470c34bbff8849b6ff6a1c7a31f81cdd80bdc813d6ddb90"} Nov 28 14:42:01 crc kubenswrapper[4817]: I1128 14:42:01.422310 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfq4b" event={"ID":"39068359-e185-4366-9378-718043d84ba2","Type":"ContainerStarted","Data":"526b21d777c9c4f6ebfb1e8c992f3b6d3fdd7d50dfdff5ac10436cf6bd6ff7e6"} Nov 28 14:42:01 crc kubenswrapper[4817]: I1128 14:42:01.422633 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:42:01 crc kubenswrapper[4817]: I1128 14:42:01.441737 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mfq4b" podStartSLOduration=6.654381617 podStartE2EDuration="13.441700469s" podCreationTimestamp="2025-11-28 14:41:48 +0000 UTC" firstStartedPulling="2025-11-28 14:41:50.01451646 +0000 UTC m=+732.602494726" lastFinishedPulling="2025-11-28 14:41:56.801835302 +0000 UTC m=+739.389813578" observedRunningTime="2025-11-28 14:42:01.44050405 +0000 UTC m=+744.028482316" watchObservedRunningTime="2025-11-28 14:42:01.441700469 +0000 UTC m=+744.029678735" Nov 28 14:42:04 crc kubenswrapper[4817]: I1128 14:42:04.869957 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:42:04 crc kubenswrapper[4817]: I1128 14:42:04.916879 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:42:09 crc kubenswrapper[4817]: I1128 14:42:09.296885 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-4r56x" Nov 28 14:42:09 crc kubenswrapper[4817]: I1128 14:42:09.872810 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mfq4b" Nov 28 14:42:09 crc kubenswrapper[4817]: I1128 14:42:09.995752 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-4wsbn" Nov 28 14:42:10 crc kubenswrapper[4817]: I1128 14:42:10.880399 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hkbsn" Nov 28 14:42:16 crc kubenswrapper[4817]: I1128 14:42:16.555086 4817 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.131164 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dvq2g"] Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.132030 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.134377 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.134450 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-lvdm6" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.139232 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.147051 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dvq2g"] Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.222637 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klhpw\" (UniqueName: \"kubernetes.io/projected/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8-kube-api-access-klhpw\") pod \"openstack-operator-index-dvq2g\" (UID: \"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8\") " pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.326745 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klhpw\" (UniqueName: \"kubernetes.io/projected/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8-kube-api-access-klhpw\") pod \"openstack-operator-index-dvq2g\" (UID: \"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8\") " pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.367416 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klhpw\" (UniqueName: \"kubernetes.io/projected/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8-kube-api-access-klhpw\") pod \"openstack-operator-index-dvq2g\" (UID: \"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8\") " pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.462494 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:17 crc kubenswrapper[4817]: I1128 14:42:17.927816 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dvq2g"] Nov 28 14:42:18 crc kubenswrapper[4817]: I1128 14:42:18.527549 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dvq2g" event={"ID":"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8","Type":"ContainerStarted","Data":"d6757afbd65d2e69011fcee69e637194f087cfc50b25538daef521108fb5adb0"} Nov 28 14:42:21 crc kubenswrapper[4817]: I1128 14:42:21.546152 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dvq2g" event={"ID":"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8","Type":"ContainerStarted","Data":"16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7"} Nov 28 14:42:21 crc kubenswrapper[4817]: I1128 14:42:21.564013 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dvq2g" podStartSLOduration=1.134941872 podStartE2EDuration="4.563992457s" podCreationTimestamp="2025-11-28 14:42:17 +0000 UTC" firstStartedPulling="2025-11-28 14:42:17.943119021 +0000 UTC m=+760.531097287" lastFinishedPulling="2025-11-28 14:42:21.372169606 +0000 UTC m=+763.960147872" observedRunningTime="2025-11-28 14:42:21.562846879 +0000 UTC m=+764.150825155" watchObservedRunningTime="2025-11-28 14:42:21.563992457 +0000 UTC m=+764.151970723" Nov 28 14:42:22 crc kubenswrapper[4817]: I1128 14:42:22.329773 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dvq2g"] Nov 28 14:42:22 crc kubenswrapper[4817]: I1128 14:42:22.931334 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mt7jk"] Nov 28 14:42:22 crc kubenswrapper[4817]: I1128 14:42:22.932983 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:22 crc kubenswrapper[4817]: I1128 14:42:22.946163 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mt7jk"] Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.009248 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5r4h\" (UniqueName: \"kubernetes.io/projected/3efb685f-68cf-4455-9e24-ac2d54d45dab-kube-api-access-t5r4h\") pod \"openstack-operator-index-mt7jk\" (UID: \"3efb685f-68cf-4455-9e24-ac2d54d45dab\") " pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.110539 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5r4h\" (UniqueName: \"kubernetes.io/projected/3efb685f-68cf-4455-9e24-ac2d54d45dab-kube-api-access-t5r4h\") pod \"openstack-operator-index-mt7jk\" (UID: \"3efb685f-68cf-4455-9e24-ac2d54d45dab\") " pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.134948 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5r4h\" (UniqueName: \"kubernetes.io/projected/3efb685f-68cf-4455-9e24-ac2d54d45dab-kube-api-access-t5r4h\") pod \"openstack-operator-index-mt7jk\" (UID: \"3efb685f-68cf-4455-9e24-ac2d54d45dab\") " pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.263264 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.558616 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-dvq2g" podUID="63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" containerName="registry-server" containerID="cri-o://16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7" gracePeriod=2 Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.716968 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mt7jk"] Nov 28 14:42:23 crc kubenswrapper[4817]: I1128 14:42:23.924298 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.027261 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klhpw\" (UniqueName: \"kubernetes.io/projected/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8-kube-api-access-klhpw\") pod \"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8\" (UID: \"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8\") " Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.032576 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8-kube-api-access-klhpw" (OuterVolumeSpecName: "kube-api-access-klhpw") pod "63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" (UID: "63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8"). InnerVolumeSpecName "kube-api-access-klhpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.045278 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.045335 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.129053 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klhpw\" (UniqueName: \"kubernetes.io/projected/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8-kube-api-access-klhpw\") on node \"crc\" DevicePath \"\"" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.566228 4817 generic.go:334] "Generic (PLEG): container finished" podID="63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" containerID="16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7" exitCode=0 Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.566299 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dvq2g" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.566320 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dvq2g" event={"ID":"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8","Type":"ContainerDied","Data":"16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7"} Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.566512 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dvq2g" event={"ID":"63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8","Type":"ContainerDied","Data":"d6757afbd65d2e69011fcee69e637194f087cfc50b25538daef521108fb5adb0"} Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.566553 4817 scope.go:117] "RemoveContainer" containerID="16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.569701 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mt7jk" event={"ID":"3efb685f-68cf-4455-9e24-ac2d54d45dab","Type":"ContainerStarted","Data":"f0ee608270d7161626781b30b635c785d0317fac83b43dc54479b759928c6757"} Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.569761 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mt7jk" event={"ID":"3efb685f-68cf-4455-9e24-ac2d54d45dab","Type":"ContainerStarted","Data":"894c144d3ffcf7f00cce6f5ea80a232eecb34475f8ab559e1b6be2e08f4d1ac7"} Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.587114 4817 scope.go:117] "RemoveContainer" containerID="16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7" Nov 28 14:42:24 crc kubenswrapper[4817]: E1128 14:42:24.588280 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7\": container with ID starting with 16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7 not found: ID does not exist" containerID="16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.588344 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7"} err="failed to get container status \"16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7\": rpc error: code = NotFound desc = could not find container \"16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7\": container with ID starting with 16f72d89c03a509df7ccadb1c92bd7ae2349dc85fdffbc0d997b31a74ccbe8a7 not found: ID does not exist" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.605327 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mt7jk" podStartSLOduration=2.562389739 podStartE2EDuration="2.605296828s" podCreationTimestamp="2025-11-28 14:42:22 +0000 UTC" firstStartedPulling="2025-11-28 14:42:23.727502357 +0000 UTC m=+766.315480643" lastFinishedPulling="2025-11-28 14:42:23.770409466 +0000 UTC m=+766.358387732" observedRunningTime="2025-11-28 14:42:24.597787402 +0000 UTC m=+767.185765688" watchObservedRunningTime="2025-11-28 14:42:24.605296828 +0000 UTC m=+767.193275104" Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.620248 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-dvq2g"] Nov 28 14:42:24 crc kubenswrapper[4817]: I1128 14:42:24.626456 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-dvq2g"] Nov 28 14:42:25 crc kubenswrapper[4817]: I1128 14:42:25.754276 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" path="/var/lib/kubelet/pods/63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8/volumes" Nov 28 14:42:33 crc kubenswrapper[4817]: I1128 14:42:33.263682 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:33 crc kubenswrapper[4817]: I1128 14:42:33.265153 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:33 crc kubenswrapper[4817]: I1128 14:42:33.288824 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:33 crc kubenswrapper[4817]: I1128 14:42:33.663890 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-mt7jk" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.787660 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc"] Nov 28 14:42:47 crc kubenswrapper[4817]: E1128 14:42:47.788445 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" containerName="registry-server" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.788457 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" containerName="registry-server" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.788569 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d8596c-6449-4bcb-9ca1-e6ab5f5a79a8" containerName="registry-server" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.789390 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.791876 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9stfs" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.795105 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc"] Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.884983 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-util\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.885032 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4vxb\" (UniqueName: \"kubernetes.io/projected/ef879694-f24b-49a3-98d2-82a71aa98d11-kube-api-access-v4vxb\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.885072 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-bundle\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.986140 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-util\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.986208 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4vxb\" (UniqueName: \"kubernetes.io/projected/ef879694-f24b-49a3-98d2-82a71aa98d11-kube-api-access-v4vxb\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.986258 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-bundle\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.986830 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-util\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:47 crc kubenswrapper[4817]: I1128 14:42:47.986833 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-bundle\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:48 crc kubenswrapper[4817]: I1128 14:42:48.011341 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4vxb\" (UniqueName: \"kubernetes.io/projected/ef879694-f24b-49a3-98d2-82a71aa98d11-kube-api-access-v4vxb\") pod \"dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:48 crc kubenswrapper[4817]: I1128 14:42:48.114139 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:48 crc kubenswrapper[4817]: I1128 14:42:48.620344 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc"] Nov 28 14:42:48 crc kubenswrapper[4817]: I1128 14:42:48.733599 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" event={"ID":"ef879694-f24b-49a3-98d2-82a71aa98d11","Type":"ContainerStarted","Data":"39f21ee8e47fd39f20d40dfbd7da27e0b607779cf915dda2091236bdbdfc26a4"} Nov 28 14:42:50 crc kubenswrapper[4817]: I1128 14:42:50.746167 4817 generic.go:334] "Generic (PLEG): container finished" podID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerID="398a5923d3a99706ac49ef0a5eba2b48ddf8349a4057ad7e1a43b9a807b0dea1" exitCode=0 Nov 28 14:42:50 crc kubenswrapper[4817]: I1128 14:42:50.746252 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" event={"ID":"ef879694-f24b-49a3-98d2-82a71aa98d11","Type":"ContainerDied","Data":"398a5923d3a99706ac49ef0a5eba2b48ddf8349a4057ad7e1a43b9a807b0dea1"} Nov 28 14:42:51 crc kubenswrapper[4817]: I1128 14:42:51.769298 4817 generic.go:334] "Generic (PLEG): container finished" podID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerID="513c1877421ce4d5496e07782637f563e760c22318a04fc1a5a176388101e35d" exitCode=0 Nov 28 14:42:51 crc kubenswrapper[4817]: I1128 14:42:51.769347 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" event={"ID":"ef879694-f24b-49a3-98d2-82a71aa98d11","Type":"ContainerDied","Data":"513c1877421ce4d5496e07782637f563e760c22318a04fc1a5a176388101e35d"} Nov 28 14:42:52 crc kubenswrapper[4817]: I1128 14:42:52.778568 4817 generic.go:334] "Generic (PLEG): container finished" podID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerID="5e37ad29d4f7cbed399493de0602b44c627715cddd1bda5620649828df7408b0" exitCode=0 Nov 28 14:42:52 crc kubenswrapper[4817]: I1128 14:42:52.778627 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" event={"ID":"ef879694-f24b-49a3-98d2-82a71aa98d11","Type":"ContainerDied","Data":"5e37ad29d4f7cbed399493de0602b44c627715cddd1bda5620649828df7408b0"} Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.045498 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.045881 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.055561 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.173999 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-bundle\") pod \"ef879694-f24b-49a3-98d2-82a71aa98d11\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.174041 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4vxb\" (UniqueName: \"kubernetes.io/projected/ef879694-f24b-49a3-98d2-82a71aa98d11-kube-api-access-v4vxb\") pod \"ef879694-f24b-49a3-98d2-82a71aa98d11\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.174164 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-util\") pod \"ef879694-f24b-49a3-98d2-82a71aa98d11\" (UID: \"ef879694-f24b-49a3-98d2-82a71aa98d11\") " Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.175707 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-bundle" (OuterVolumeSpecName: "bundle") pod "ef879694-f24b-49a3-98d2-82a71aa98d11" (UID: "ef879694-f24b-49a3-98d2-82a71aa98d11"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.187238 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-util" (OuterVolumeSpecName: "util") pod "ef879694-f24b-49a3-98d2-82a71aa98d11" (UID: "ef879694-f24b-49a3-98d2-82a71aa98d11"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.191245 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef879694-f24b-49a3-98d2-82a71aa98d11-kube-api-access-v4vxb" (OuterVolumeSpecName: "kube-api-access-v4vxb") pod "ef879694-f24b-49a3-98d2-82a71aa98d11" (UID: "ef879694-f24b-49a3-98d2-82a71aa98d11"). InnerVolumeSpecName "kube-api-access-v4vxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.276150 4817 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-util\") on node \"crc\" DevicePath \"\"" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.276194 4817 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef879694-f24b-49a3-98d2-82a71aa98d11-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.276208 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4vxb\" (UniqueName: \"kubernetes.io/projected/ef879694-f24b-49a3-98d2-82a71aa98d11-kube-api-access-v4vxb\") on node \"crc\" DevicePath \"\"" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.792774 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" event={"ID":"ef879694-f24b-49a3-98d2-82a71aa98d11","Type":"ContainerDied","Data":"39f21ee8e47fd39f20d40dfbd7da27e0b607779cf915dda2091236bdbdfc26a4"} Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.792823 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39f21ee8e47fd39f20d40dfbd7da27e0b607779cf915dda2091236bdbdfc26a4" Nov 28 14:42:54 crc kubenswrapper[4817]: I1128 14:42:54.792862 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.141249 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh"] Nov 28 14:43:00 crc kubenswrapper[4817]: E1128 14:43:00.142378 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="pull" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.142394 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="pull" Nov 28 14:43:00 crc kubenswrapper[4817]: E1128 14:43:00.142414 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="extract" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.142422 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="extract" Nov 28 14:43:00 crc kubenswrapper[4817]: E1128 14:43:00.142440 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="util" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.142448 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="util" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.142737 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef879694-f24b-49a3-98d2-82a71aa98d11" containerName="extract" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.143400 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:00 crc kubenswrapper[4817]: W1128 14:43:00.152559 4817 reflector.go:561] object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-d94f2": failed to list *v1.Secret: secrets "openstack-operator-controller-operator-dockercfg-d94f2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Nov 28 14:43:00 crc kubenswrapper[4817]: E1128 14:43:00.152820 4817 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"openstack-operator-controller-operator-dockercfg-d94f2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openstack-operator-controller-operator-dockercfg-d94f2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.236786 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh"] Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.264580 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfhv8\" (UniqueName: \"kubernetes.io/projected/cb0ec512-1d2d-4487-8d7e-c412ea2b4da6-kube-api-access-pfhv8\") pod \"openstack-operator-controller-operator-56ffd7cc6d-rw9vh\" (UID: \"cb0ec512-1d2d-4487-8d7e-c412ea2b4da6\") " pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.365813 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfhv8\" (UniqueName: \"kubernetes.io/projected/cb0ec512-1d2d-4487-8d7e-c412ea2b4da6-kube-api-access-pfhv8\") pod \"openstack-operator-controller-operator-56ffd7cc6d-rw9vh\" (UID: \"cb0ec512-1d2d-4487-8d7e-c412ea2b4da6\") " pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:00 crc kubenswrapper[4817]: I1128 14:43:00.389190 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfhv8\" (UniqueName: \"kubernetes.io/projected/cb0ec512-1d2d-4487-8d7e-c412ea2b4da6-kube-api-access-pfhv8\") pod \"openstack-operator-controller-operator-56ffd7cc6d-rw9vh\" (UID: \"cb0ec512-1d2d-4487-8d7e-c412ea2b4da6\") " pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:01 crc kubenswrapper[4817]: I1128 14:43:01.483700 4817 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" secret="" err="failed to sync secret cache: timed out waiting for the condition" Nov 28 14:43:01 crc kubenswrapper[4817]: I1128 14:43:01.483863 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:01 crc kubenswrapper[4817]: I1128 14:43:01.681485 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh"] Nov 28 14:43:01 crc kubenswrapper[4817]: I1128 14:43:01.702107 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-d94f2" Nov 28 14:43:01 crc kubenswrapper[4817]: I1128 14:43:01.836302 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" event={"ID":"cb0ec512-1d2d-4487-8d7e-c412ea2b4da6","Type":"ContainerStarted","Data":"fd4be4362b3e5df3f9252472d80ffc07902f2fec617b9cb9acf1ac86e3d0ca3d"} Nov 28 14:43:06 crc kubenswrapper[4817]: I1128 14:43:06.865834 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" event={"ID":"cb0ec512-1d2d-4487-8d7e-c412ea2b4da6","Type":"ContainerStarted","Data":"38e4636569f4f7e7a6fb92d72d6f312ff29ce51466f5c02cdb6bbee0f8b7e901"} Nov 28 14:43:06 crc kubenswrapper[4817]: I1128 14:43:06.866503 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:06 crc kubenswrapper[4817]: I1128 14:43:06.898270 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" podStartSLOduration=2.8162257410000002 podStartE2EDuration="6.898251309s" podCreationTimestamp="2025-11-28 14:43:00 +0000 UTC" firstStartedPulling="2025-11-28 14:43:01.691660505 +0000 UTC m=+804.279638771" lastFinishedPulling="2025-11-28 14:43:05.773686073 +0000 UTC m=+808.361664339" observedRunningTime="2025-11-28 14:43:06.896199209 +0000 UTC m=+809.484177475" watchObservedRunningTime="2025-11-28 14:43:06.898251309 +0000 UTC m=+809.486229575" Nov 28 14:43:11 crc kubenswrapper[4817]: I1128 14:43:11.487889 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-56ffd7cc6d-rw9vh" Nov 28 14:43:24 crc kubenswrapper[4817]: I1128 14:43:24.045047 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:43:24 crc kubenswrapper[4817]: I1128 14:43:24.045655 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:43:24 crc kubenswrapper[4817]: I1128 14:43:24.045708 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:43:24 crc kubenswrapper[4817]: I1128 14:43:24.046387 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e638df1b21b62724d5463d8dfc0e4379271e7a660caed8cd6069eeac89d71b58"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:43:24 crc kubenswrapper[4817]: I1128 14:43:24.046448 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://e638df1b21b62724d5463d8dfc0e4379271e7a660caed8cd6069eeac89d71b58" gracePeriod=600 Nov 28 14:43:26 crc kubenswrapper[4817]: I1128 14:43:26.093550 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="e638df1b21b62724d5463d8dfc0e4379271e7a660caed8cd6069eeac89d71b58" exitCode=0 Nov 28 14:43:26 crc kubenswrapper[4817]: I1128 14:43:26.093639 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"e638df1b21b62724d5463d8dfc0e4379271e7a660caed8cd6069eeac89d71b58"} Nov 28 14:43:26 crc kubenswrapper[4817]: I1128 14:43:26.094837 4817 scope.go:117] "RemoveContainer" containerID="93f0ad5fc8ef7c51ce6b53865c3f234a722e3cdac3537164a439b62ef13ee382" Nov 28 14:43:27 crc kubenswrapper[4817]: I1128 14:43:27.101280 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"f352602e89e72622a4635f595006f3c005b7dadf542d0db30e6fcfcc7b955afc"} Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.419626 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.421178 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.426152 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.427062 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.427132 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xmstp" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.431657 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6j6d5" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.435977 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.448282 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-8cjxb"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.449550 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.453323 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-f8xgv" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.454786 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.465466 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.466544 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.471250 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-k5vwv" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.483603 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-8cjxb"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.505344 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.509939 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.511074 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.526104 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-hnhwv" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.536343 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.537302 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.544168 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mdstn" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.544674 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.554385 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.560831 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.562037 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.565359 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kst7m\" (UniqueName: \"kubernetes.io/projected/b0544c31-6ad6-4a76-84da-c180f89ba5ee-kube-api-access-kst7m\") pod \"cinder-operator-controller-manager-6b7f75547b-75m4b\" (UID: \"b0544c31-6ad6-4a76-84da-c180f89ba5ee\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.565397 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjg47\" (UniqueName: \"kubernetes.io/projected/59388928-9001-4abd-9361-d4045b4d027b-kube-api-access-hjg47\") pod \"barbican-operator-controller-manager-7b64f4fb85-8lb8j\" (UID: \"59388928-9001-4abd-9361-d4045b4d027b\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.565443 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhsk2\" (UniqueName: \"kubernetes.io/projected/025353e3-a121-42b5-b283-727b6eb24b7c-kube-api-access-mhsk2\") pod \"designate-operator-controller-manager-955677c94-8cjxb\" (UID: \"025353e3-a121-42b5-b283-727b6eb24b7c\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.565470 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj9qq\" (UniqueName: \"kubernetes.io/projected/29cc111e-478d-48c6-a89c-4050114b5d37-kube-api-access-rj9qq\") pod \"glance-operator-controller-manager-589cbd6b5b-4fj4z\" (UID: \"29cc111e-478d-48c6-a89c-4050114b5d37\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.569155 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.569238 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-97ff5" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.570858 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.582736 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.584029 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.588062 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-skb4l" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.588500 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.600860 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.605244 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.609341 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-c5t9t" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.629568 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.647261 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.648444 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.652031 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-zzsd9" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.660022 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.664376 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.665306 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.666409 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj5sp\" (UniqueName: \"kubernetes.io/projected/694e1cb6-aa3a-4909-b220-99d42dbf1e49-kube-api-access-kj5sp\") pod \"keystone-operator-controller-manager-7b4567c7cf-4phnt\" (UID: \"694e1cb6-aa3a-4909-b220-99d42dbf1e49\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.666455 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qbv5\" (UniqueName: \"kubernetes.io/projected/e04c0207-7d18-437a-81a3-51ca263e492b-kube-api-access-7qbv5\") pod \"horizon-operator-controller-manager-5d494799bf-j2vsj\" (UID: \"e04c0207-7d18-437a-81a3-51ca263e492b\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.666498 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmgnm\" (UniqueName: \"kubernetes.io/projected/a55aea0f-52dc-4403-8067-2372b24cdf83-kube-api-access-bmgnm\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.666589 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.666811 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kst7m\" (UniqueName: \"kubernetes.io/projected/b0544c31-6ad6-4a76-84da-c180f89ba5ee-kube-api-access-kst7m\") pod \"cinder-operator-controller-manager-6b7f75547b-75m4b\" (UID: \"b0544c31-6ad6-4a76-84da-c180f89ba5ee\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.666847 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjg47\" (UniqueName: \"kubernetes.io/projected/59388928-9001-4abd-9361-d4045b4d027b-kube-api-access-hjg47\") pod \"barbican-operator-controller-manager-7b64f4fb85-8lb8j\" (UID: \"59388928-9001-4abd-9361-d4045b4d027b\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.667007 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj2ct\" (UniqueName: \"kubernetes.io/projected/02f94412-5d1b-401d-a9fa-211df97a2710-kube-api-access-xj2ct\") pod \"heat-operator-controller-manager-5b77f656f-k7tbk\" (UID: \"02f94412-5d1b-401d-a9fa-211df97a2710\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.667036 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhsk2\" (UniqueName: \"kubernetes.io/projected/025353e3-a121-42b5-b283-727b6eb24b7c-kube-api-access-mhsk2\") pod \"designate-operator-controller-manager-955677c94-8cjxb\" (UID: \"025353e3-a121-42b5-b283-727b6eb24b7c\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.667073 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj9qq\" (UniqueName: \"kubernetes.io/projected/29cc111e-478d-48c6-a89c-4050114b5d37-kube-api-access-rj9qq\") pod \"glance-operator-controller-manager-589cbd6b5b-4fj4z\" (UID: \"29cc111e-478d-48c6-a89c-4050114b5d37\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.667109 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltq6t\" (UniqueName: \"kubernetes.io/projected/4598ac96-dead-47cc-b063-22d3197689a4-kube-api-access-ltq6t\") pod \"ironic-operator-controller-manager-67cb4dc6d4-vh4rl\" (UID: \"4598ac96-dead-47cc-b063-22d3197689a4\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.672036 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-rmkzs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.695236 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kst7m\" (UniqueName: \"kubernetes.io/projected/b0544c31-6ad6-4a76-84da-c180f89ba5ee-kube-api-access-kst7m\") pod \"cinder-operator-controller-manager-6b7f75547b-75m4b\" (UID: \"b0544c31-6ad6-4a76-84da-c180f89ba5ee\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.695301 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.696183 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.698465 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qppb6" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.699815 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.700361 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj9qq\" (UniqueName: \"kubernetes.io/projected/29cc111e-478d-48c6-a89c-4050114b5d37-kube-api-access-rj9qq\") pod \"glance-operator-controller-manager-589cbd6b5b-4fj4z\" (UID: \"29cc111e-478d-48c6-a89c-4050114b5d37\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.712525 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.713812 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.715985 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-z4pqh" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.716532 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjg47\" (UniqueName: \"kubernetes.io/projected/59388928-9001-4abd-9361-d4045b4d027b-kube-api-access-hjg47\") pod \"barbican-operator-controller-manager-7b64f4fb85-8lb8j\" (UID: \"59388928-9001-4abd-9361-d4045b4d027b\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.717222 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhsk2\" (UniqueName: \"kubernetes.io/projected/025353e3-a121-42b5-b283-727b6eb24b7c-kube-api-access-mhsk2\") pod \"designate-operator-controller-manager-955677c94-8cjxb\" (UID: \"025353e3-a121-42b5-b283-727b6eb24b7c\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.755146 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.756434 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.764899 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.757335 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.805573 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmgnm\" (UniqueName: \"kubernetes.io/projected/a55aea0f-52dc-4403-8067-2372b24cdf83-kube-api-access-bmgnm\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.805649 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.805813 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pgd4\" (UniqueName: \"kubernetes.io/projected/19220366-8f0a-4108-9093-964f54a4642f-kube-api-access-5pgd4\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-87bqn\" (UID: \"19220366-8f0a-4108-9093-964f54a4642f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.805875 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.805943 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj2ct\" (UniqueName: \"kubernetes.io/projected/02f94412-5d1b-401d-a9fa-211df97a2710-kube-api-access-xj2ct\") pod \"heat-operator-controller-manager-5b77f656f-k7tbk\" (UID: \"02f94412-5d1b-401d-a9fa-211df97a2710\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.805999 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gr8\" (UniqueName: \"kubernetes.io/projected/2f81ca1e-831f-4a15-9ea9-a81364b534fd-kube-api-access-57gr8\") pod \"neutron-operator-controller-manager-6fdcddb789-pmw8p\" (UID: \"2f81ca1e-831f-4a15-9ea9-a81364b534fd\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.806055 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltq6t\" (UniqueName: \"kubernetes.io/projected/4598ac96-dead-47cc-b063-22d3197689a4-kube-api-access-ltq6t\") pod \"ironic-operator-controller-manager-67cb4dc6d4-vh4rl\" (UID: \"4598ac96-dead-47cc-b063-22d3197689a4\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.806109 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4l8h\" (UniqueName: \"kubernetes.io/projected/be902d34-5e5e-48be-9a6b-d9e53661f367-kube-api-access-h4l8h\") pod \"manila-operator-controller-manager-5d499bf58b-tbssr\" (UID: \"be902d34-5e5e-48be-9a6b-d9e53661f367\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.806143 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj5sp\" (UniqueName: \"kubernetes.io/projected/694e1cb6-aa3a-4909-b220-99d42dbf1e49-kube-api-access-kj5sp\") pod \"keystone-operator-controller-manager-7b4567c7cf-4phnt\" (UID: \"694e1cb6-aa3a-4909-b220-99d42dbf1e49\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.806166 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qbv5\" (UniqueName: \"kubernetes.io/projected/e04c0207-7d18-437a-81a3-51ca263e492b-kube-api-access-7qbv5\") pod \"horizon-operator-controller-manager-5d494799bf-j2vsj\" (UID: \"e04c0207-7d18-437a-81a3-51ca263e492b\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.806538 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:43:35 crc kubenswrapper[4817]: E1128 14:43:35.807055 4817 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:35 crc kubenswrapper[4817]: E1128 14:43:35.807103 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert podName:a55aea0f-52dc-4403-8067-2372b24cdf83 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:36.307083827 +0000 UTC m=+838.895062093 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert") pod "infra-operator-controller-manager-57548d458d-w9qqs" (UID: "a55aea0f-52dc-4403-8067-2372b24cdf83") : secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.810838 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.824834 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-mvst5" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.840340 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qbv5\" (UniqueName: \"kubernetes.io/projected/e04c0207-7d18-437a-81a3-51ca263e492b-kube-api-access-7qbv5\") pod \"horizon-operator-controller-manager-5d494799bf-j2vsj\" (UID: \"e04c0207-7d18-437a-81a3-51ca263e492b\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.849652 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj2ct\" (UniqueName: \"kubernetes.io/projected/02f94412-5d1b-401d-a9fa-211df97a2710-kube-api-access-xj2ct\") pod \"heat-operator-controller-manager-5b77f656f-k7tbk\" (UID: \"02f94412-5d1b-401d-a9fa-211df97a2710\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.851389 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmgnm\" (UniqueName: \"kubernetes.io/projected/a55aea0f-52dc-4403-8067-2372b24cdf83-kube-api-access-bmgnm\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.856924 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.856936 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.859636 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj5sp\" (UniqueName: \"kubernetes.io/projected/694e1cb6-aa3a-4909-b220-99d42dbf1e49-kube-api-access-kj5sp\") pod \"keystone-operator-controller-manager-7b4567c7cf-4phnt\" (UID: \"694e1cb6-aa3a-4909-b220-99d42dbf1e49\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.860244 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltq6t\" (UniqueName: \"kubernetes.io/projected/4598ac96-dead-47cc-b063-22d3197689a4-kube-api-access-ltq6t\") pod \"ironic-operator-controller-manager-67cb4dc6d4-vh4rl\" (UID: \"4598ac96-dead-47cc-b063-22d3197689a4\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.871445 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.888767 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.891726 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.895144 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.895389 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-76jt4" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.906991 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.912284 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwtjz\" (UniqueName: \"kubernetes.io/projected/18412c56-76ce-48a3-81f2-3cbc4d090df5-kube-api-access-qwtjz\") pod \"octavia-operator-controller-manager-64cdc6ff96-gv25c\" (UID: \"18412c56-76ce-48a3-81f2-3cbc4d090df5\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.912348 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.912361 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4l8h\" (UniqueName: \"kubernetes.io/projected/be902d34-5e5e-48be-9a6b-d9e53661f367-kube-api-access-h4l8h\") pod \"manila-operator-controller-manager-5d499bf58b-tbssr\" (UID: \"be902d34-5e5e-48be-9a6b-d9e53661f367\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.912470 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pgd4\" (UniqueName: \"kubernetes.io/projected/19220366-8f0a-4108-9093-964f54a4642f-kube-api-access-5pgd4\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-87bqn\" (UID: \"19220366-8f0a-4108-9093-964f54a4642f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.912590 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkr4s\" (UniqueName: \"kubernetes.io/projected/abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29-kube-api-access-zkr4s\") pod \"nova-operator-controller-manager-79556f57fc-wzw2n\" (UID: \"abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.912652 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57gr8\" (UniqueName: \"kubernetes.io/projected/2f81ca1e-831f-4a15-9ea9-a81364b534fd-kube-api-access-57gr8\") pod \"neutron-operator-controller-manager-6fdcddb789-pmw8p\" (UID: \"2f81ca1e-831f-4a15-9ea9-a81364b534fd\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.928837 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.929610 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.932061 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5p8mf" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.939133 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.940520 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.941747 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4l8h\" (UniqueName: \"kubernetes.io/projected/be902d34-5e5e-48be-9a6b-d9e53661f367-kube-api-access-h4l8h\") pod \"manila-operator-controller-manager-5d499bf58b-tbssr\" (UID: \"be902d34-5e5e-48be-9a6b-d9e53661f367\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.943275 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.943905 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-g98f8" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.953653 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pgd4\" (UniqueName: \"kubernetes.io/projected/19220366-8f0a-4108-9093-964f54a4642f-kube-api-access-5pgd4\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-87bqn\" (UID: \"19220366-8f0a-4108-9093-964f54a4642f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.958753 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.963181 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gr8\" (UniqueName: \"kubernetes.io/projected/2f81ca1e-831f-4a15-9ea9-a81364b534fd-kube-api-access-57gr8\") pod \"neutron-operator-controller-manager-6fdcddb789-pmw8p\" (UID: \"2f81ca1e-831f-4a15-9ea9-a81364b534fd\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.965713 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.971820 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-n96rm"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.972070 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.972957 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.976946 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4b5zq" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.981998 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-n96rm"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.986869 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n"] Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.987903 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.988249 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:43:35 crc kubenswrapper[4817]: I1128 14:43:35.996602 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-dcmlx" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.001479 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.013645 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkr4s\" (UniqueName: \"kubernetes.io/projected/abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29-kube-api-access-zkr4s\") pod \"nova-operator-controller-manager-79556f57fc-wzw2n\" (UID: \"abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.013739 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwtjz\" (UniqueName: \"kubernetes.io/projected/18412c56-76ce-48a3-81f2-3cbc4d090df5-kube-api-access-qwtjz\") pod \"octavia-operator-controller-manager-64cdc6ff96-gv25c\" (UID: \"18412c56-76ce-48a3-81f2-3cbc4d090df5\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.013778 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m2dm\" (UniqueName: \"kubernetes.io/projected/35934979-2374-4a86-b9db-7effb7624f40-kube-api-access-2m2dm\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.013810 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.013827 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqcn2\" (UniqueName: \"kubernetes.io/projected/6ec7b27a-79e7-402c-8b73-091c9c20a1db-kube-api-access-nqcn2\") pod \"ovn-operator-controller-manager-56897c768d-s52pg\" (UID: \"6ec7b27a-79e7-402c-8b73-091c9c20a1db\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.037116 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkr4s\" (UniqueName: \"kubernetes.io/projected/abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29-kube-api-access-zkr4s\") pod \"nova-operator-controller-manager-79556f57fc-wzw2n\" (UID: \"abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.043976 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwtjz\" (UniqueName: \"kubernetes.io/projected/18412c56-76ce-48a3-81f2-3cbc4d090df5-kube-api-access-qwtjz\") pod \"octavia-operator-controller-manager-64cdc6ff96-gv25c\" (UID: \"18412c56-76ce-48a3-81f2-3cbc4d090df5\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.075085 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.080564 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.082619 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-h9fxh" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.114523 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.116973 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m2dm\" (UniqueName: \"kubernetes.io/projected/35934979-2374-4a86-b9db-7effb7624f40-kube-api-access-2m2dm\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.117038 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.117063 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqcn2\" (UniqueName: \"kubernetes.io/projected/6ec7b27a-79e7-402c-8b73-091c9c20a1db-kube-api-access-nqcn2\") pod \"ovn-operator-controller-manager-56897c768d-s52pg\" (UID: \"6ec7b27a-79e7-402c-8b73-091c9c20a1db\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.117701 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24ph9\" (UniqueName: \"kubernetes.io/projected/cafcb9ae-aa1a-490e-b27d-656bf4b8508d-kube-api-access-24ph9\") pod \"telemetry-operator-controller-manager-5b7b8977cb-jkx7n\" (UID: \"cafcb9ae-aa1a-490e-b27d-656bf4b8508d\") " pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.117759 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69fcj\" (UniqueName: \"kubernetes.io/projected/a023b80a-0d1c-47c6-b904-014d659e559a-kube-api-access-69fcj\") pod \"swift-operator-controller-manager-d77b94747-n96rm\" (UID: \"a023b80a-0d1c-47c6-b904-014d659e559a\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.117812 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cxh4\" (UniqueName: \"kubernetes.io/projected/5343b3a1-d2c2-4b26-806b-a7661275d20f-kube-api-access-9cxh4\") pod \"placement-operator-controller-manager-57988cc5b5-8fqs4\" (UID: \"5343b3a1-d2c2-4b26-806b-a7661275d20f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.118390 4817 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.118468 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert podName:35934979-2374-4a86-b9db-7effb7624f40 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:36.618448817 +0000 UTC m=+839.206427083 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" (UID: "35934979-2374-4a86-b9db-7effb7624f40") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.141430 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m2dm\" (UniqueName: \"kubernetes.io/projected/35934979-2374-4a86-b9db-7effb7624f40-kube-api-access-2m2dm\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.145032 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.145471 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.145581 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqcn2\" (UniqueName: \"kubernetes.io/projected/6ec7b27a-79e7-402c-8b73-091c9c20a1db-kube-api-access-nqcn2\") pod \"ovn-operator-controller-manager-56897c768d-s52pg\" (UID: \"6ec7b27a-79e7-402c-8b73-091c9c20a1db\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.162679 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.171050 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.173197 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.175881 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.176633 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-t8pkw" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.177139 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.191833 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.192733 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.195473 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.196991 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-cjwmz" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.199951 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.218375 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.219924 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24ph9\" (UniqueName: \"kubernetes.io/projected/cafcb9ae-aa1a-490e-b27d-656bf4b8508d-kube-api-access-24ph9\") pod \"telemetry-operator-controller-manager-5b7b8977cb-jkx7n\" (UID: \"cafcb9ae-aa1a-490e-b27d-656bf4b8508d\") " pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.219964 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69fcj\" (UniqueName: \"kubernetes.io/projected/a023b80a-0d1c-47c6-b904-014d659e559a-kube-api-access-69fcj\") pod \"swift-operator-controller-manager-d77b94747-n96rm\" (UID: \"a023b80a-0d1c-47c6-b904-014d659e559a\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.220014 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cxh4\" (UniqueName: \"kubernetes.io/projected/5343b3a1-d2c2-4b26-806b-a7661275d20f-kube-api-access-9cxh4\") pod \"placement-operator-controller-manager-57988cc5b5-8fqs4\" (UID: \"5343b3a1-d2c2-4b26-806b-a7661275d20f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.220048 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v6p2\" (UniqueName: \"kubernetes.io/projected/e8b4fc65-1b95-41fb-a95e-6e102c452691-kube-api-access-2v6p2\") pod \"test-operator-controller-manager-5cd6c7f4c8-xpvwd\" (UID: \"e8b4fc65-1b95-41fb-a95e-6e102c452691\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.249144 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69fcj\" (UniqueName: \"kubernetes.io/projected/a023b80a-0d1c-47c6-b904-014d659e559a-kube-api-access-69fcj\") pod \"swift-operator-controller-manager-d77b94747-n96rm\" (UID: \"a023b80a-0d1c-47c6-b904-014d659e559a\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.250767 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cxh4\" (UniqueName: \"kubernetes.io/projected/5343b3a1-d2c2-4b26-806b-a7661275d20f-kube-api-access-9cxh4\") pod \"placement-operator-controller-manager-57988cc5b5-8fqs4\" (UID: \"5343b3a1-d2c2-4b26-806b-a7661275d20f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.261614 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.268325 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.272311 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-g9c7h" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.272644 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24ph9\" (UniqueName: \"kubernetes.io/projected/cafcb9ae-aa1a-490e-b27d-656bf4b8508d-kube-api-access-24ph9\") pod \"telemetry-operator-controller-manager-5b7b8977cb-jkx7n\" (UID: \"cafcb9ae-aa1a-490e-b27d-656bf4b8508d\") " pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.276695 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.283153 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.305344 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.328365 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hmml\" (UniqueName: \"kubernetes.io/projected/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-kube-api-access-4hmml\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.328402 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v6p2\" (UniqueName: \"kubernetes.io/projected/e8b4fc65-1b95-41fb-a95e-6e102c452691-kube-api-access-2v6p2\") pod \"test-operator-controller-manager-5cd6c7f4c8-xpvwd\" (UID: \"e8b4fc65-1b95-41fb-a95e-6e102c452691\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.328442 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrhtl\" (UniqueName: \"kubernetes.io/projected/77766fad-8504-4053-a3f6-78591e568ce5-kube-api-access-mrhtl\") pod \"watcher-operator-controller-manager-656dcb59d4-26v9h\" (UID: \"77766fad-8504-4053-a3f6-78591e568ce5\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.328477 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.328497 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.328528 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.328952 4817 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.328991 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert podName:a55aea0f-52dc-4403-8067-2372b24cdf83 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:37.328976989 +0000 UTC m=+839.916955245 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert") pod "infra-operator-controller-manager-57548d458d-w9qqs" (UID: "a55aea0f-52dc-4403-8067-2372b24cdf83") : secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.341441 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.348500 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v6p2\" (UniqueName: \"kubernetes.io/projected/e8b4fc65-1b95-41fb-a95e-6e102c452691-kube-api-access-2v6p2\") pod \"test-operator-controller-manager-5cd6c7f4c8-xpvwd\" (UID: \"e8b4fc65-1b95-41fb-a95e-6e102c452691\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.359166 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.359668 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.430560 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgszs\" (UniqueName: \"kubernetes.io/projected/d9e5f208-4552-48ff-af77-901fa7c76e42-kube-api-access-hgszs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jr75r\" (UID: \"d9e5f208-4552-48ff-af77-901fa7c76e42\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.431010 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hmml\" (UniqueName: \"kubernetes.io/projected/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-kube-api-access-4hmml\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.431060 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrhtl\" (UniqueName: \"kubernetes.io/projected/77766fad-8504-4053-a3f6-78591e568ce5-kube-api-access-mrhtl\") pod \"watcher-operator-controller-manager-656dcb59d4-26v9h\" (UID: \"77766fad-8504-4053-a3f6-78591e568ce5\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.431106 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.431156 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.431566 4817 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.431632 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:36.931601621 +0000 UTC m=+839.519579887 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "metrics-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.434629 4817 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.434706 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:36.934684397 +0000 UTC m=+839.522662663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.455973 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.456263 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrhtl\" (UniqueName: \"kubernetes.io/projected/77766fad-8504-4053-a3f6-78591e568ce5-kube-api-access-mrhtl\") pod \"watcher-operator-controller-manager-656dcb59d4-26v9h\" (UID: \"77766fad-8504-4053-a3f6-78591e568ce5\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.459986 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hmml\" (UniqueName: \"kubernetes.io/projected/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-kube-api-access-4hmml\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.546592 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgszs\" (UniqueName: \"kubernetes.io/projected/d9e5f208-4552-48ff-af77-901fa7c76e42-kube-api-access-hgszs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jr75r\" (UID: \"d9e5f208-4552-48ff-af77-901fa7c76e42\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.559626 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.578569 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgszs\" (UniqueName: \"kubernetes.io/projected/d9e5f208-4552-48ff-af77-901fa7c76e42-kube-api-access-hgszs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jr75r\" (UID: \"d9e5f208-4552-48ff-af77-901fa7c76e42\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.609246 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.625242 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.633642 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj"] Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.649614 4817 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.649699 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert podName:35934979-2374-4a86-b9db-7effb7624f40 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:37.649680719 +0000 UTC m=+840.237658985 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" (UID: "35934979-2374-4a86-b9db-7effb7624f40") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.650168 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.871470 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.902671 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt"] Nov 28 14:43:36 crc kubenswrapper[4817]: W1128 14:43:36.902907 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod694e1cb6_aa3a_4909_b220_99d42dbf1e49.slice/crio-ac095b25c5d5217c8985498d3fb41730f371bc3c2f47558983e1b9b4aee068df WatchSource:0}: Error finding container ac095b25c5d5217c8985498d3fb41730f371bc3c2f47558983e1b9b4aee068df: Status 404 returned error can't find the container with id ac095b25c5d5217c8985498d3fb41730f371bc3c2f47558983e1b9b4aee068df Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.907284 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.912858 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.916925 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-8cjxb"] Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.953795 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.954017 4817 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.954306 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:37.954287852 +0000 UTC m=+840.542266118 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "metrics-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.954462 4817 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: E1128 14:43:36.954490 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:37.954482757 +0000 UTC m=+840.542461023 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "webhook-server-cert" not found Nov 28 14:43:36 crc kubenswrapper[4817]: I1128 14:43:36.955140 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.046449 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.170884 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.189113 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.220537 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.223669 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" event={"ID":"025353e3-a121-42b5-b283-727b6eb24b7c","Type":"ContainerStarted","Data":"62b4da60b794657922edc74201b7802f508e6bdcd04e31d7c13b0be3b40b8974"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.227640 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" event={"ID":"e04c0207-7d18-437a-81a3-51ca263e492b","Type":"ContainerStarted","Data":"77f164e6c6a5fe25f4bfceac1283b6d457695725f78a71a02fa205008e8b117a"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.243347 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" event={"ID":"b0544c31-6ad6-4a76-84da-c180f89ba5ee","Type":"ContainerStarted","Data":"692d975b32334566109ee8b930e192e471f1559db02f5ec3c5952244ec866b6c"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.248021 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" event={"ID":"694e1cb6-aa3a-4909-b220-99d42dbf1e49","Type":"ContainerStarted","Data":"ac095b25c5d5217c8985498d3fb41730f371bc3c2f47558983e1b9b4aee068df"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.249216 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" event={"ID":"29cc111e-478d-48c6-a89c-4050114b5d37","Type":"ContainerStarted","Data":"0da3feeae3234ac45dcf3798234a583c437926a60142e8624a2d5c1caed4a480"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.250021 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" event={"ID":"59388928-9001-4abd-9361-d4045b4d027b","Type":"ContainerStarted","Data":"68af6d8e9b2cfbceba0b044fcfb34d365c81ad6e1e64ae40cfdaad2bdf02d147"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.250882 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" event={"ID":"be902d34-5e5e-48be-9a6b-d9e53661f367","Type":"ContainerStarted","Data":"597a7d4c26bd621d2313f30f56616e2591439e5cb56d7a753ac50462bfd9c83b"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.251625 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" event={"ID":"2f81ca1e-831f-4a15-9ea9-a81364b534fd","Type":"ContainerStarted","Data":"1a7de745e6a9f8d5ca347bcf04b0ad16791bdb2663bb7a667907062f54017e6f"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.252379 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" event={"ID":"4598ac96-dead-47cc-b063-22d3197689a4","Type":"ContainerStarted","Data":"ce12a764b43c82d9836431a15b0a289bb15cae5792b36595eef46372c0900534"} Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.277900 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.284023 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.304773 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c"] Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.306420 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2v6p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-xpvwd_openstack-operators(e8b4fc65-1b95-41fb-a95e-6e102c452691): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.315136 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2v6p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-xpvwd_openstack-operators(e8b4fc65-1b95-41fb-a95e-6e102c452691): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.316228 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" podUID="e8b4fc65-1b95-41fb-a95e-6e102c452691" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.317142 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qwtjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-gv25c_openstack-operators(18412c56-76ce-48a3-81f2-3cbc4d090df5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: W1128 14:43:37.318877 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ec7b27a_79e7_402c_8b73_091c9c20a1db.slice/crio-af327c951c57af7844ae8b2d3c2f9060dc06db18a8f361b53a23cd1e53732888 WatchSource:0}: Error finding container af327c951c57af7844ae8b2d3c2f9060dc06db18a8f361b53a23cd1e53732888: Status 404 returned error can't find the container with id af327c951c57af7844ae8b2d3c2f9060dc06db18a8f361b53a23cd1e53732888 Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.322627 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg"] Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.326918 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-69fcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-n96rm_openstack-operators(a023b80a-0d1c-47c6-b904-014d659e559a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.327039 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hgszs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jr75r_openstack-operators(d9e5f208-4552-48ff-af77-901fa7c76e42): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.328127 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" podUID="d9e5f208-4552-48ff-af77-901fa7c76e42" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.328261 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nqcn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-s52pg_openstack-operators(6ec7b27a-79e7-402c-8b73-091c9c20a1db): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.328686 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-69fcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-n96rm_openstack-operators(a023b80a-0d1c-47c6-b904-014d659e559a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.328747 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h"] Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.329942 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" podUID="a023b80a-0d1c-47c6-b904-014d659e559a" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.332570 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nqcn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-s52pg_openstack-operators(6ec7b27a-79e7-402c-8b73-091c9c20a1db): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: W1128 14:43:37.332610 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77766fad_8504_4053_a3f6_78591e568ce5.slice/crio-96547d349ffac5e561441a02059b0481dcdf30b648abb27afbe5d39a11953b4a WatchSource:0}: Error finding container 96547d349ffac5e561441a02059b0481dcdf30b648abb27afbe5d39a11953b4a: Status 404 returned error can't find the container with id 96547d349ffac5e561441a02059b0481dcdf30b648abb27afbe5d39a11953b4a Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.334533 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" podUID="6ec7b27a-79e7-402c-8b73-091c9c20a1db" Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.336661 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-n96rm"] Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.343314 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r"] Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.344320 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mrhtl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-26v9h_openstack-operators(77766fad-8504-4053-a3f6-78591e568ce5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.346074 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mrhtl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-26v9h_openstack-operators(77766fad-8504-4053-a3f6-78591e568ce5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.347814 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" podUID="77766fad-8504-4053-a3f6-78591e568ce5" Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.366322 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.366472 4817 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.366515 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert podName:a55aea0f-52dc-4403-8067-2372b24cdf83 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:39.366503909 +0000 UTC m=+841.954482175 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert") pod "infra-operator-controller-manager-57548d458d-w9qqs" (UID: "a55aea0f-52dc-4403-8067-2372b24cdf83") : secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.462882 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n"] Nov 28 14:43:37 crc kubenswrapper[4817]: W1128 14:43:37.475101 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcafcb9ae_aa1a_490e_b27d_656bf4b8508d.slice/crio-245dc4eb3e38ff2b3b94ee92dc31edf3f0406469393f109d25acd7bd6ea5a503 WatchSource:0}: Error finding container 245dc4eb3e38ff2b3b94ee92dc31edf3f0406469393f109d25acd7bd6ea5a503: Status 404 returned error can't find the container with id 245dc4eb3e38ff2b3b94ee92dc31edf3f0406469393f109d25acd7bd6ea5a503 Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.476508 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.2:5001/openstack-k8s-operators/telemetry-operator:a56cff847472bbc2ff74c1f159f60d5390d3c1bf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-24ph9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5b7b8977cb-jkx7n_openstack-operators(cafcb9ae-aa1a-490e-b27d-656bf4b8508d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.478767 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-24ph9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5b7b8977cb-jkx7n_openstack-operators(cafcb9ae-aa1a-490e-b27d-656bf4b8508d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.480029 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" podUID="cafcb9ae-aa1a-490e-b27d-656bf4b8508d" Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.673246 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.673399 4817 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.673456 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert podName:35934979-2374-4a86-b9db-7effb7624f40 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:39.673437479 +0000 UTC m=+842.261415745 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" (UID: "35934979-2374-4a86-b9db-7effb7624f40") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.985494 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:37 crc kubenswrapper[4817]: I1128 14:43:37.985896 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.986097 4817 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.986154 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:39.986137141 +0000 UTC m=+842.574115407 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "webhook-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.986492 4817 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 14:43:37 crc kubenswrapper[4817]: E1128 14:43:37.986526 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:39.98651848 +0000 UTC m=+842.574496746 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "metrics-server-cert" not found Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.302018 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" event={"ID":"6ec7b27a-79e7-402c-8b73-091c9c20a1db","Type":"ContainerStarted","Data":"af327c951c57af7844ae8b2d3c2f9060dc06db18a8f361b53a23cd1e53732888"} Nov 28 14:43:38 crc kubenswrapper[4817]: E1128 14:43:38.306876 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" podUID="6ec7b27a-79e7-402c-8b73-091c9c20a1db" Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.308293 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" event={"ID":"abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29","Type":"ContainerStarted","Data":"b8e3e028381f7dc8771c205319e5f322b51a468f7b150bdd52dd2446e8d5bef6"} Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.309901 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" event={"ID":"77766fad-8504-4053-a3f6-78591e568ce5","Type":"ContainerStarted","Data":"96547d349ffac5e561441a02059b0481dcdf30b648abb27afbe5d39a11953b4a"} Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.311912 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" event={"ID":"19220366-8f0a-4108-9093-964f54a4642f","Type":"ContainerStarted","Data":"0e74a9aba52648746041a2926ac64c3e10cfdc5473cb2cfcd34e217ff901d3d8"} Nov 28 14:43:38 crc kubenswrapper[4817]: E1128 14:43:38.312286 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" podUID="77766fad-8504-4053-a3f6-78591e568ce5" Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.314109 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" event={"ID":"a023b80a-0d1c-47c6-b904-014d659e559a","Type":"ContainerStarted","Data":"56c0eff4af33fbefaa4c4494c53649b9706d2fc46cb87c957d7523afd2f68757"} Nov 28 14:43:38 crc kubenswrapper[4817]: E1128 14:43:38.319275 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" podUID="a023b80a-0d1c-47c6-b904-014d659e559a" Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.323029 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" event={"ID":"d9e5f208-4552-48ff-af77-901fa7c76e42","Type":"ContainerStarted","Data":"caf168dac86d40c0dcd24c39b39ac52f5c15d40a2bf8fed23890c501b9c3df9e"} Nov 28 14:43:38 crc kubenswrapper[4817]: E1128 14:43:38.325367 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" podUID="d9e5f208-4552-48ff-af77-901fa7c76e42" Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.327465 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" event={"ID":"18412c56-76ce-48a3-81f2-3cbc4d090df5","Type":"ContainerStarted","Data":"80db24e68b60f13a88abb604c2c99ba7199ecf3d8ca507fe00d339f9b5326546"} Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.329072 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" event={"ID":"02f94412-5d1b-401d-a9fa-211df97a2710","Type":"ContainerStarted","Data":"2b33a6346863e63a2fbc6b764d7c0b71bd22c158f291f74069c790e35eb11331"} Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.330572 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" event={"ID":"cafcb9ae-aa1a-490e-b27d-656bf4b8508d","Type":"ContainerStarted","Data":"245dc4eb3e38ff2b3b94ee92dc31edf3f0406469393f109d25acd7bd6ea5a503"} Nov 28 14:43:38 crc kubenswrapper[4817]: E1128 14:43:38.332326 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.2:5001/openstack-k8s-operators/telemetry-operator:a56cff847472bbc2ff74c1f159f60d5390d3c1bf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" podUID="cafcb9ae-aa1a-490e-b27d-656bf4b8508d" Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.337184 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" event={"ID":"e8b4fc65-1b95-41fb-a95e-6e102c452691","Type":"ContainerStarted","Data":"f881484aff2b0809809a9daad3e75348299b0d4237eb22f78249c256b717ca1d"} Nov 28 14:43:38 crc kubenswrapper[4817]: E1128 14:43:38.343075 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" podUID="e8b4fc65-1b95-41fb-a95e-6e102c452691" Nov 28 14:43:38 crc kubenswrapper[4817]: I1128 14:43:38.346173 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" event={"ID":"5343b3a1-d2c2-4b26-806b-a7661275d20f","Type":"ContainerStarted","Data":"53d07ca768a4bdfaf865381ac8e96d5926f6244b921d693231072ed5da7ccc12"} Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.373062 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" podUID="d9e5f208-4552-48ff-af77-901fa7c76e42" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.373139 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" podUID="e8b4fc65-1b95-41fb-a95e-6e102c452691" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.373466 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" podUID="a023b80a-0d1c-47c6-b904-014d659e559a" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.373530 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" podUID="77766fad-8504-4053-a3f6-78591e568ce5" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.373680 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" podUID="6ec7b27a-79e7-402c-8b73-091c9c20a1db" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.373792 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.2:5001/openstack-k8s-operators/telemetry-operator:a56cff847472bbc2ff74c1f159f60d5390d3c1bf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" podUID="cafcb9ae-aa1a-490e-b27d-656bf4b8508d" Nov 28 14:43:39 crc kubenswrapper[4817]: I1128 14:43:39.403503 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.403656 4817 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.403707 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert podName:a55aea0f-52dc-4403-8067-2372b24cdf83 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:43.403690384 +0000 UTC m=+845.991668650 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert") pod "infra-operator-controller-manager-57548d458d-w9qqs" (UID: "a55aea0f-52dc-4403-8067-2372b24cdf83") : secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:39 crc kubenswrapper[4817]: I1128 14:43:39.709813 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.710213 4817 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:39 crc kubenswrapper[4817]: E1128 14:43:39.710263 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert podName:35934979-2374-4a86-b9db-7effb7624f40 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:43.710247515 +0000 UTC m=+846.298225781 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" (UID: "35934979-2374-4a86-b9db-7effb7624f40") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:40 crc kubenswrapper[4817]: I1128 14:43:40.019675 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:40 crc kubenswrapper[4817]: I1128 14:43:40.019821 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:40 crc kubenswrapper[4817]: E1128 14:43:40.019874 4817 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 14:43:40 crc kubenswrapper[4817]: E1128 14:43:40.019964 4817 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 14:43:40 crc kubenswrapper[4817]: E1128 14:43:40.020017 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:44.019938923 +0000 UTC m=+846.607917189 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "webhook-server-cert" not found Nov 28 14:43:40 crc kubenswrapper[4817]: E1128 14:43:40.020039 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:44.020030455 +0000 UTC m=+846.608008721 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "metrics-server-cert" not found Nov 28 14:43:43 crc kubenswrapper[4817]: I1128 14:43:43.408033 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:43 crc kubenswrapper[4817]: E1128 14:43:43.408227 4817 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:43 crc kubenswrapper[4817]: E1128 14:43:43.408801 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert podName:a55aea0f-52dc-4403-8067-2372b24cdf83 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:51.408776526 +0000 UTC m=+853.996754792 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert") pod "infra-operator-controller-manager-57548d458d-w9qqs" (UID: "a55aea0f-52dc-4403-8067-2372b24cdf83") : secret "infra-operator-webhook-server-cert" not found Nov 28 14:43:43 crc kubenswrapper[4817]: I1128 14:43:43.713064 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:43 crc kubenswrapper[4817]: E1128 14:43:43.713334 4817 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:43 crc kubenswrapper[4817]: E1128 14:43:43.713496 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert podName:35934979-2374-4a86-b9db-7effb7624f40 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:51.713462317 +0000 UTC m=+854.301440583 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" (UID: "35934979-2374-4a86-b9db-7effb7624f40") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 14:43:44 crc kubenswrapper[4817]: I1128 14:43:44.119835 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:44 crc kubenswrapper[4817]: I1128 14:43:44.119907 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:44 crc kubenswrapper[4817]: E1128 14:43:44.120013 4817 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 14:43:44 crc kubenswrapper[4817]: E1128 14:43:44.120185 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:52.120158059 +0000 UTC m=+854.708136365 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "metrics-server-cert" not found Nov 28 14:43:44 crc kubenswrapper[4817]: E1128 14:43:44.120027 4817 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 14:43:44 crc kubenswrapper[4817]: E1128 14:43:44.120275 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs podName:c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5 nodeName:}" failed. No retries permitted until 2025-11-28 14:43:52.120256201 +0000 UTC m=+854.708234467 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs") pod "openstack-operator-controller-manager-64d6f76c7b-5smh2" (UID: "c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5") : secret "webhook-server-cert" not found Nov 28 14:43:50 crc kubenswrapper[4817]: E1128 14:43:50.312038 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2" Nov 28 14:43:50 crc kubenswrapper[4817]: E1128 14:43:50.314810 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rj9qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-589cbd6b5b-4fj4z_openstack-operators(29cc111e-478d-48c6-a89c-4050114b5d37): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:50 crc kubenswrapper[4817]: E1128 14:43:50.865264 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:d65dbfc956e9cf376f3c48fc3a0942cb7306b5164f898c40d1efca106df81db7" Nov 28 14:43:50 crc kubenswrapper[4817]: E1128 14:43:50.865463 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:d65dbfc956e9cf376f3c48fc3a0942cb7306b5164f898c40d1efca106df81db7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ltq6t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-67cb4dc6d4-vh4rl_openstack-operators(4598ac96-dead-47cc-b063-22d3197689a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:51 crc kubenswrapper[4817]: E1128 14:43:51.379937 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:ca332e48d07f932e470177e48dba9332848a1d14c857cff6f9bfb1adc1998482" Nov 28 14:43:51 crc kubenswrapper[4817]: E1128 14:43:51.380132 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:ca332e48d07f932e470177e48dba9332848a1d14c857cff6f9bfb1adc1998482,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kst7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6b7f75547b-75m4b_openstack-operators(b0544c31-6ad6-4a76-84da-c180f89ba5ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.428014 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.447188 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a55aea0f-52dc-4403-8067-2372b24cdf83-cert\") pod \"infra-operator-controller-manager-57548d458d-w9qqs\" (UID: \"a55aea0f-52dc-4403-8067-2372b24cdf83\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.481564 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-97ff5" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.490539 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.733229 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.736762 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/35934979-2374-4a86-b9db-7effb7624f40-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm\" (UID: \"35934979-2374-4a86-b9db-7effb7624f40\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.868731 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-76jt4" Nov 28 14:43:51 crc kubenswrapper[4817]: I1128 14:43:51.878116 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:43:52 crc kubenswrapper[4817]: I1128 14:43:52.137924 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:52 crc kubenswrapper[4817]: I1128 14:43:52.137979 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:52 crc kubenswrapper[4817]: I1128 14:43:52.143306 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-webhook-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:52 crc kubenswrapper[4817]: I1128 14:43:52.153855 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5-metrics-certs\") pod \"openstack-operator-controller-manager-64d6f76c7b-5smh2\" (UID: \"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5\") " pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:52 crc kubenswrapper[4817]: I1128 14:43:52.200928 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-cjwmz" Nov 28 14:43:52 crc kubenswrapper[4817]: I1128 14:43:52.209412 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:43:56 crc kubenswrapper[4817]: E1128 14:43:56.773630 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677" Nov 28 14:43:56 crc kubenswrapper[4817]: E1128 14:43:56.774492 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xj2ct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5b77f656f-k7tbk_openstack-operators(02f94412-5d1b-401d-a9fa-211df97a2710): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:57 crc kubenswrapper[4817]: E1128 14:43:57.438164 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 28 14:43:57 crc kubenswrapper[4817]: E1128 14:43:57.438597 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zkr4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-wzw2n_openstack-operators(abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:58 crc kubenswrapper[4817]: E1128 14:43:58.004379 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711" Nov 28 14:43:58 crc kubenswrapper[4817]: E1128 14:43:58.004562 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kj5sp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-4phnt_openstack-operators(694e1cb6-aa3a-4909-b220-99d42dbf1e49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:58 crc kubenswrapper[4817]: E1128 14:43:58.436311 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423" Nov 28 14:43:58 crc kubenswrapper[4817]: E1128 14:43:58.436476 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9cxh4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-8fqs4_openstack-operators(5343b3a1-d2c2-4b26-806b-a7661275d20f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:58 crc kubenswrapper[4817]: E1128 14:43:58.811549 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c" Nov 28 14:43:58 crc kubenswrapper[4817]: E1128 14:43:58.811831 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-57gr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-pmw8p_openstack-operators(2f81ca1e-831f-4a15-9ea9-a81364b534fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:43:59 crc kubenswrapper[4817]: E1128 14:43:59.231234 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2" Nov 28 14:43:59 crc kubenswrapper[4817]: E1128 14:43:59.231758 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5pgd4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-87bqn_openstack-operators(19220366-8f0a-4108-9093-964f54a4642f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:44:05 crc kubenswrapper[4817]: I1128 14:44:05.848088 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2"] Nov 28 14:44:05 crc kubenswrapper[4817]: I1128 14:44:05.854438 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm"] Nov 28 14:44:05 crc kubenswrapper[4817]: I1128 14:44:05.956237 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs"] Nov 28 14:44:05 crc kubenswrapper[4817]: W1128 14:44:05.964874 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6203ca7_1ea0_44ae_8c4f_cc3637ae0af5.slice/crio-e880a91e91bb8e47274e3e6ec814be4088069c383288c07e7ace51ad27134f59 WatchSource:0}: Error finding container e880a91e91bb8e47274e3e6ec814be4088069c383288c07e7ace51ad27134f59: Status 404 returned error can't find the container with id e880a91e91bb8e47274e3e6ec814be4088069c383288c07e7ace51ad27134f59 Nov 28 14:44:05 crc kubenswrapper[4817]: W1128 14:44:05.971283 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35934979_2374_4a86_b9db_7effb7624f40.slice/crio-79f5e1d5b3c7a9c30f55c88915c9638294138238f8e9a41f46e690fd5c54fdb4 WatchSource:0}: Error finding container 79f5e1d5b3c7a9c30f55c88915c9638294138238f8e9a41f46e690fd5c54fdb4: Status 404 returned error can't find the container with id 79f5e1d5b3c7a9c30f55c88915c9638294138238f8e9a41f46e690fd5c54fdb4 Nov 28 14:44:05 crc kubenswrapper[4817]: W1128 14:44:05.984830 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda55aea0f_52dc_4403_8067_2372b24cdf83.slice/crio-a1e55ef7494c2d8cf07900cd101c9df7b4ffd3e3dbdb155d3fffa11522790cbf WatchSource:0}: Error finding container a1e55ef7494c2d8cf07900cd101c9df7b4ffd3e3dbdb155d3fffa11522790cbf: Status 404 returned error can't find the container with id a1e55ef7494c2d8cf07900cd101c9df7b4ffd3e3dbdb155d3fffa11522790cbf Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.568611 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" event={"ID":"e04c0207-7d18-437a-81a3-51ca263e492b","Type":"ContainerStarted","Data":"50d8725ab4757f82f1280ca605ab2eb22cb1b63cdcbd6ceb00ac966dbaa54297"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.570452 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" event={"ID":"a55aea0f-52dc-4403-8067-2372b24cdf83","Type":"ContainerStarted","Data":"a1e55ef7494c2d8cf07900cd101c9df7b4ffd3e3dbdb155d3fffa11522790cbf"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.572369 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" event={"ID":"025353e3-a121-42b5-b283-727b6eb24b7c","Type":"ContainerStarted","Data":"585214b6263af959622181ce5f6bbce45b9f6d701f578355a29dc39abfdf4d0c"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.574261 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" event={"ID":"59388928-9001-4abd-9361-d4045b4d027b","Type":"ContainerStarted","Data":"57ab7e0f6537b727aca55ea310ac406703e1a166583f742d44c0c02c3202ecf7"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.583475 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" event={"ID":"77766fad-8504-4053-a3f6-78591e568ce5","Type":"ContainerStarted","Data":"188cb0c18725edbbada2aa347b895454a94572d398fe60c11fd8afbecd407ded"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.590128 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" event={"ID":"35934979-2374-4a86-b9db-7effb7624f40","Type":"ContainerStarted","Data":"79f5e1d5b3c7a9c30f55c88915c9638294138238f8e9a41f46e690fd5c54fdb4"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.592743 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" event={"ID":"a023b80a-0d1c-47c6-b904-014d659e559a","Type":"ContainerStarted","Data":"92e1713b568832b54d045be3acc065504b689f2f3aee4b2d916dafb3a0e18f48"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.596264 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" event={"ID":"e8b4fc65-1b95-41fb-a95e-6e102c452691","Type":"ContainerStarted","Data":"3471d701e9bca160ef6555dc6e9faa7b910f3301ed013dc825ae7714b9b6bfcf"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.600751 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" event={"ID":"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5","Type":"ContainerStarted","Data":"e880a91e91bb8e47274e3e6ec814be4088069c383288c07e7ace51ad27134f59"} Nov 28 14:44:06 crc kubenswrapper[4817]: I1128 14:44:06.605104 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" event={"ID":"be902d34-5e5e-48be-9a6b-d9e53661f367","Type":"ContainerStarted","Data":"c736dd16c40810753906983da80d859609b262a8f9824a31f771b5eae5235a87"} Nov 28 14:44:07 crc kubenswrapper[4817]: I1128 14:44:07.623188 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" event={"ID":"6ec7b27a-79e7-402c-8b73-091c9c20a1db","Type":"ContainerStarted","Data":"e274bfeafcd1f35333b07acb2aa471b0a82b3e76c8f4e9f6e710c465b2a266e9"} Nov 28 14:44:07 crc kubenswrapper[4817]: I1128 14:44:07.651652 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" event={"ID":"c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5","Type":"ContainerStarted","Data":"60bfce4e287a1f7f65dcb4d859149377f02e7b31a36739fc5af223ade3c6c7dc"} Nov 28 14:44:07 crc kubenswrapper[4817]: I1128 14:44:07.763023 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" podStartSLOduration=31.763005223 podStartE2EDuration="31.763005223s" podCreationTimestamp="2025-11-28 14:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:44:07.67936397 +0000 UTC m=+870.267342236" watchObservedRunningTime="2025-11-28 14:44:07.763005223 +0000 UTC m=+870.350983489" Nov 28 14:44:08 crc kubenswrapper[4817]: I1128 14:44:08.658121 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" event={"ID":"d9e5f208-4552-48ff-af77-901fa7c76e42","Type":"ContainerStarted","Data":"53ac5e058d2184bf20c8f03d77f4ff97f1fad67d7108896cbbdbd41de28d5d74"} Nov 28 14:44:08 crc kubenswrapper[4817]: I1128 14:44:08.659901 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" event={"ID":"cafcb9ae-aa1a-490e-b27d-656bf4b8508d","Type":"ContainerStarted","Data":"4bf4cc3cda328fddaaab2256b26908c83bdbf6f1713e7b4ff5bbaa52620d054b"} Nov 28 14:44:08 crc kubenswrapper[4817]: I1128 14:44:08.660092 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:44:08 crc kubenswrapper[4817]: I1128 14:44:08.678380 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jr75r" podStartSLOduration=3.913621902 podStartE2EDuration="32.678363907s" podCreationTimestamp="2025-11-28 14:43:36 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.326937173 +0000 UTC m=+839.914915439" lastFinishedPulling="2025-11-28 14:44:06.091679178 +0000 UTC m=+868.679657444" observedRunningTime="2025-11-28 14:44:08.676416449 +0000 UTC m=+871.264394715" watchObservedRunningTime="2025-11-28 14:44:08.678363907 +0000 UTC m=+871.266342173" Nov 28 14:44:09 crc kubenswrapper[4817]: E1128 14:44:09.867113 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" podUID="b0544c31-6ad6-4a76-84da-c180f89ba5ee" Nov 28 14:44:09 crc kubenswrapper[4817]: E1128 14:44:09.915078 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" podUID="02f94412-5d1b-401d-a9fa-211df97a2710" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.086358 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" podUID="29cc111e-478d-48c6-a89c-4050114b5d37" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.208639 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" podUID="4598ac96-dead-47cc-b063-22d3197689a4" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.230256 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" podUID="694e1cb6-aa3a-4909-b220-99d42dbf1e49" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.280136 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" podUID="5343b3a1-d2c2-4b26-806b-a7661275d20f" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.315931 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" podUID="2f81ca1e-831f-4a15-9ea9-a81364b534fd" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.320324 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" podUID="abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.322405 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" podUID="19220366-8f0a-4108-9093-964f54a4642f" Nov 28 14:44:10 crc kubenswrapper[4817]: E1128 14:44:10.335320 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" podUID="18412c56-76ce-48a3-81f2-3cbc4d090df5" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.685066 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" event={"ID":"e04c0207-7d18-437a-81a3-51ca263e492b","Type":"ContainerStarted","Data":"a9aa3d94e06991ba2b8b51ef0cbaaf43b5dd01a3de45c4853f036c48162a3b93"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.685424 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.686432 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" event={"ID":"19220366-8f0a-4108-9093-964f54a4642f","Type":"ContainerStarted","Data":"1c1dfa22db60dc5f70614851ad8d650e18a0627e4c69ab171ce83b9571021b4b"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.687956 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.688003 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" event={"ID":"a55aea0f-52dc-4403-8067-2372b24cdf83","Type":"ContainerStarted","Data":"e8fa54c1430eaef83296202fb130436c68691f515b65f37ddac1516456fa566e"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.688044 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" event={"ID":"a55aea0f-52dc-4403-8067-2372b24cdf83","Type":"ContainerStarted","Data":"05ce4be60c62ff1c02c583b9fef793f90537cd029fe65e0e96457cdf6c0eb76c"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.688093 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.690210 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" event={"ID":"02f94412-5d1b-401d-a9fa-211df97a2710","Type":"ContainerStarted","Data":"9bd3f1b839ec8515790d33ae7aa7fb71c1f453f737423653abb799637efe439b"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.692216 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" event={"ID":"6ec7b27a-79e7-402c-8b73-091c9c20a1db","Type":"ContainerStarted","Data":"c34074c84c27ec8fce76a9b8dff9387122ac7f12d2f9917e0c279b6bd360d400"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.692551 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.693769 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" event={"ID":"29cc111e-478d-48c6-a89c-4050114b5d37","Type":"ContainerStarted","Data":"11df64c2fb83f54fb67fee5d7924320952459a8b607fbdeab6828d806db6fae6"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.696755 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" event={"ID":"59388928-9001-4abd-9361-d4045b4d027b","Type":"ContainerStarted","Data":"da298e5e49c47fd2bb0f9932d2b78475a4a41f0f007846b6353da8b6f13e92f2"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.697204 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.698360 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" event={"ID":"abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29","Type":"ContainerStarted","Data":"336a659d88b4e6bab66eb80ecc3a9a26db11d7fceb09aa4e2bf2ce4a9b3e3446"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.704861 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" event={"ID":"cafcb9ae-aa1a-490e-b27d-656bf4b8508d","Type":"ContainerStarted","Data":"42dc6a0a2aa73805aca0bae7962e06f18c9babaa51f35cbc87f263d2fa372f46"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.705414 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.709675 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" podStartSLOduration=3.052469499 podStartE2EDuration="35.709659701s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.718461036 +0000 UTC m=+839.306439302" lastFinishedPulling="2025-11-28 14:44:09.375651238 +0000 UTC m=+871.963629504" observedRunningTime="2025-11-28 14:44:10.704127043 +0000 UTC m=+873.292105309" watchObservedRunningTime="2025-11-28 14:44:10.709659701 +0000 UTC m=+873.297637967" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.721355 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" event={"ID":"694e1cb6-aa3a-4909-b220-99d42dbf1e49","Type":"ContainerStarted","Data":"395a01857c62160c706696cee1408506a801e6d6e349333d4626d96cb5bab555"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.723362 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" event={"ID":"025353e3-a121-42b5-b283-727b6eb24b7c","Type":"ContainerStarted","Data":"d4acb452e592f20df357e4b2fcdfbc9bd4a2b376f97d76c5eaf81ca57d5f12d4"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.723504 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.730237 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" event={"ID":"5343b3a1-d2c2-4b26-806b-a7661275d20f","Type":"ContainerStarted","Data":"8da45d5b4da39e3a15b0682c11a1367033a5c1694327b85cf8f732006762ba4b"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.742989 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" event={"ID":"77766fad-8504-4053-a3f6-78591e568ce5","Type":"ContainerStarted","Data":"472bf9463a2854946e8d04b544e1a051d3e8a541f6eac0429510d9f711c05b35"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.743584 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.767460 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" event={"ID":"a023b80a-0d1c-47c6-b904-014d659e559a","Type":"ContainerStarted","Data":"13198922ff7f5c7d43503115eeef37d6ff0dcace923406194ec04e8ae041c8a7"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.767581 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.783434 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" event={"ID":"e8b4fc65-1b95-41fb-a95e-6e102c452691","Type":"ContainerStarted","Data":"6e3cb6969fa4690e7acd4b766a2d4d5617420dbbf8945ff7e34389d7111da3f1"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.784141 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.793054 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" event={"ID":"18412c56-76ce-48a3-81f2-3cbc4d090df5","Type":"ContainerStarted","Data":"af9aaf12beea18bcd7a87b969ecbe26e2e590a5037d6a94b3e5e8695139cb4f5"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.795778 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" podStartSLOduration=2.7351460789999997 podStartE2EDuration="35.795764586s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.412286244 +0000 UTC m=+839.000264510" lastFinishedPulling="2025-11-28 14:44:09.472904751 +0000 UTC m=+872.060883017" observedRunningTime="2025-11-28 14:44:10.763242216 +0000 UTC m=+873.351220492" watchObservedRunningTime="2025-11-28 14:44:10.795764586 +0000 UTC m=+873.383742852" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.811313 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" event={"ID":"b0544c31-6ad6-4a76-84da-c180f89ba5ee","Type":"ContainerStarted","Data":"7140eebff33f067c011867cea03631bd61c83c6c2bd857f4ae31934151ce866b"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.814273 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" podStartSLOduration=3.700597483 podStartE2EDuration="35.814260667s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.328195474 +0000 UTC m=+839.916173740" lastFinishedPulling="2025-11-28 14:44:09.441858658 +0000 UTC m=+872.029836924" observedRunningTime="2025-11-28 14:44:10.813897068 +0000 UTC m=+873.401875354" watchObservedRunningTime="2025-11-28 14:44:10.814260667 +0000 UTC m=+873.402238933" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.831880 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" event={"ID":"4598ac96-dead-47cc-b063-22d3197689a4","Type":"ContainerStarted","Data":"39eeb2c9fd6a637ffb189cee7f21ef780092b5a3a66436676e28b9bf0bca5339"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.883258 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" event={"ID":"35934979-2374-4a86-b9db-7effb7624f40","Type":"ContainerStarted","Data":"e082ec9f15c02162a2f5ca9b14bd3803d3245a58a6cb39ba154a88736e43171d"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.883324 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.883346 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" event={"ID":"35934979-2374-4a86-b9db-7effb7624f40","Type":"ContainerStarted","Data":"0a96ed7b274ba640c331399a8e0f6e907c948cea065b881a34d9316ce888fb7c"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.896929 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" podStartSLOduration=32.497049548 podStartE2EDuration="35.896909366s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:44:05.986795985 +0000 UTC m=+868.574774251" lastFinishedPulling="2025-11-28 14:44:09.386655803 +0000 UTC m=+871.974634069" observedRunningTime="2025-11-28 14:44:10.890207399 +0000 UTC m=+873.478185665" watchObservedRunningTime="2025-11-28 14:44:10.896909366 +0000 UTC m=+873.484887712" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.897057 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" event={"ID":"be902d34-5e5e-48be-9a6b-d9e53661f367","Type":"ContainerStarted","Data":"5bb4010d78891f6740fea1787eeb457215d596dfb9bbc066d800972f76c90c16"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.898069 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.932960 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" event={"ID":"2f81ca1e-831f-4a15-9ea9-a81364b534fd","Type":"ContainerStarted","Data":"d8c3bbdb4023b1ba9267d13bc1564df9b51e7ac6633101d735d1c7db0cb2e1ba"} Nov 28 14:44:10 crc kubenswrapper[4817]: I1128 14:44:10.981360 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" podStartSLOduration=32.560683863 podStartE2EDuration="35.981342489s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:44:05.9765842 +0000 UTC m=+868.564562466" lastFinishedPulling="2025-11-28 14:44:09.397242826 +0000 UTC m=+871.985221092" observedRunningTime="2025-11-28 14:44:10.980525139 +0000 UTC m=+873.568503415" watchObservedRunningTime="2025-11-28 14:44:10.981342489 +0000 UTC m=+873.569320755" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.062826 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" podStartSLOduration=7.970749066 podStartE2EDuration="36.062805589s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.476387129 +0000 UTC m=+840.064365395" lastFinishedPulling="2025-11-28 14:44:05.568443652 +0000 UTC m=+868.156421918" observedRunningTime="2025-11-28 14:44:11.055542558 +0000 UTC m=+873.643520824" watchObservedRunningTime="2025-11-28 14:44:11.062805589 +0000 UTC m=+873.650783855" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.082811 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" podStartSLOduration=3.656816363 podStartE2EDuration="36.082790936s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.965250213 +0000 UTC m=+839.553228479" lastFinishedPulling="2025-11-28 14:44:09.391224786 +0000 UTC m=+871.979203052" observedRunningTime="2025-11-28 14:44:11.082019597 +0000 UTC m=+873.669997863" watchObservedRunningTime="2025-11-28 14:44:11.082790936 +0000 UTC m=+873.670769202" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.158829 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" podStartSLOduration=4.014667082 podStartE2EDuration="36.15881113s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.306219312 +0000 UTC m=+839.894197568" lastFinishedPulling="2025-11-28 14:44:09.45036334 +0000 UTC m=+872.038341616" observedRunningTime="2025-11-28 14:44:11.156170185 +0000 UTC m=+873.744148471" watchObservedRunningTime="2025-11-28 14:44:11.15881113 +0000 UTC m=+873.746789396" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.182115 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" podStartSLOduration=3.069350876 podStartE2EDuration="35.18209677s" podCreationTimestamp="2025-11-28 14:43:36 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.344192439 +0000 UTC m=+839.932170705" lastFinishedPulling="2025-11-28 14:44:09.456938333 +0000 UTC m=+872.044916599" observedRunningTime="2025-11-28 14:44:11.180362077 +0000 UTC m=+873.768340343" watchObservedRunningTime="2025-11-28 14:44:11.18209677 +0000 UTC m=+873.770075036" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.196747 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" podStartSLOduration=4.098422044 podStartE2EDuration="36.196728595s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.32681769 +0000 UTC m=+839.914795956" lastFinishedPulling="2025-11-28 14:44:09.425124231 +0000 UTC m=+872.013102507" observedRunningTime="2025-11-28 14:44:11.193510885 +0000 UTC m=+873.781489161" watchObservedRunningTime="2025-11-28 14:44:11.196728595 +0000 UTC m=+873.784706861" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.209976 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" podStartSLOduration=3.690866944 podStartE2EDuration="36.209956614s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.949915194 +0000 UTC m=+839.537893450" lastFinishedPulling="2025-11-28 14:44:09.469004864 +0000 UTC m=+872.056983120" observedRunningTime="2025-11-28 14:44:11.208016096 +0000 UTC m=+873.795994362" watchObservedRunningTime="2025-11-28 14:44:11.209956614 +0000 UTC m=+873.797934890" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.951435 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" event={"ID":"abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29","Type":"ContainerStarted","Data":"533a5f3f57823c70fc3b370bbe294c99ddc8b79a263db813de04daa49ed1ef93"} Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.951800 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.971243 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" event={"ID":"694e1cb6-aa3a-4909-b220-99d42dbf1e49","Type":"ContainerStarted","Data":"7afeb55eb6da30d8c24b0830dd9377ce388578b945916155a7a40d3eee854b6a"} Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.972052 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.977313 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" podStartSLOduration=2.91302678 podStartE2EDuration="36.977296871s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.223642776 +0000 UTC m=+839.811621042" lastFinishedPulling="2025-11-28 14:44:11.287912867 +0000 UTC m=+873.875891133" observedRunningTime="2025-11-28 14:44:11.972470911 +0000 UTC m=+874.560449177" watchObservedRunningTime="2025-11-28 14:44:11.977296871 +0000 UTC m=+874.565275137" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.978580 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" event={"ID":"02f94412-5d1b-401d-a9fa-211df97a2710","Type":"ContainerStarted","Data":"83fb22b03b72e5d87e37a79cf0e7cd4e5a121bf874e38f41c2b7657eca022b58"} Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.979168 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.980830 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" event={"ID":"5343b3a1-d2c2-4b26-806b-a7661275d20f","Type":"ContainerStarted","Data":"f010956e9d8fc105679e462b482838472feb36de360e651401d0c9cf62526760"} Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.989446 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-26v9h" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.989845 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-8lb8j" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.991808 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-n96rm" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.991843 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-8cjxb" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.991867 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-tbssr" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.992978 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-s52pg" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.993074 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-xpvwd" Nov 28 14:44:11 crc kubenswrapper[4817]: I1128 14:44:11.993130 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-j2vsj" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.003005 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" podStartSLOduration=2.56180713 podStartE2EDuration="37.002982411s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.904857083 +0000 UTC m=+839.492835349" lastFinishedPulling="2025-11-28 14:44:11.346032364 +0000 UTC m=+873.934010630" observedRunningTime="2025-11-28 14:44:11.995668848 +0000 UTC m=+874.583647114" watchObservedRunningTime="2025-11-28 14:44:12.002982411 +0000 UTC m=+874.590960677" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.025131 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" podStartSLOduration=2.899370863 podStartE2EDuration="37.025110602s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.290879864 +0000 UTC m=+839.878858130" lastFinishedPulling="2025-11-28 14:44:11.416619603 +0000 UTC m=+874.004597869" observedRunningTime="2025-11-28 14:44:12.013253537 +0000 UTC m=+874.601231793" watchObservedRunningTime="2025-11-28 14:44:12.025110602 +0000 UTC m=+874.613088868" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.225612 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-64d6f76c7b-5smh2" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.291347 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" podStartSLOduration=3.226221772 podStartE2EDuration="37.291316574s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.223784249 +0000 UTC m=+839.811762515" lastFinishedPulling="2025-11-28 14:44:11.288879051 +0000 UTC m=+873.876857317" observedRunningTime="2025-11-28 14:44:12.25825158 +0000 UTC m=+874.846229846" watchObservedRunningTime="2025-11-28 14:44:12.291316574 +0000 UTC m=+874.879294860" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.988549 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" event={"ID":"19220366-8f0a-4108-9093-964f54a4642f","Type":"ContainerStarted","Data":"a4004893327dc771a1780956d19e1a46b7dde745bc089c0d5b9952bffa7d5e90"} Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.988689 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.991160 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" event={"ID":"2f81ca1e-831f-4a15-9ea9-a81364b534fd","Type":"ContainerStarted","Data":"5d7b0ed2309c613fdd4b0ed0b7b69cae1d8914c010c4b7c06ebb5f7a0ef21bd5"} Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.991447 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.993961 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" event={"ID":"29cc111e-478d-48c6-a89c-4050114b5d37","Type":"ContainerStarted","Data":"6cd233ffe4f72ba091ddd5d625f68949a2fa2c9bfaa5057df5e50fdd18dfd55c"} Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.994086 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.995996 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" event={"ID":"b0544c31-6ad6-4a76-84da-c180f89ba5ee","Type":"ContainerStarted","Data":"ebc7430b4589cee7b8640fb39fc5fc62a759ae3a848a248bf745fddd8767e2bb"} Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.996095 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.998099 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" event={"ID":"4598ac96-dead-47cc-b063-22d3197689a4","Type":"ContainerStarted","Data":"70793f18162b58cefdaa7e1aa976de72ddd0157d05a75ce5af557affec0ae049"} Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.998192 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:44:12 crc kubenswrapper[4817]: I1128 14:44:12.998368 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:44:13 crc kubenswrapper[4817]: I1128 14:44:13.011238 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" podStartSLOduration=3.647512298 podStartE2EDuration="38.011223538s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.218240062 +0000 UTC m=+839.806218328" lastFinishedPulling="2025-11-28 14:44:11.581951302 +0000 UTC m=+874.169929568" observedRunningTime="2025-11-28 14:44:13.006828719 +0000 UTC m=+875.594806985" watchObservedRunningTime="2025-11-28 14:44:13.011223538 +0000 UTC m=+875.599201804" Nov 28 14:44:13 crc kubenswrapper[4817]: I1128 14:44:13.024856 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" podStartSLOduration=3.562912866 podStartE2EDuration="38.024838768s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.119920997 +0000 UTC m=+839.707899263" lastFinishedPulling="2025-11-28 14:44:11.581846899 +0000 UTC m=+874.169825165" observedRunningTime="2025-11-28 14:44:13.021514775 +0000 UTC m=+875.609493041" watchObservedRunningTime="2025-11-28 14:44:13.024838768 +0000 UTC m=+875.612817034" Nov 28 14:44:13 crc kubenswrapper[4817]: I1128 14:44:13.048424 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" podStartSLOduration=3.383109936 podStartE2EDuration="38.048406685s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.920277273 +0000 UTC m=+839.508255539" lastFinishedPulling="2025-11-28 14:44:11.585574022 +0000 UTC m=+874.173552288" observedRunningTime="2025-11-28 14:44:13.045971284 +0000 UTC m=+875.633949550" watchObservedRunningTime="2025-11-28 14:44:13.048406685 +0000 UTC m=+875.636384961" Nov 28 14:44:13 crc kubenswrapper[4817]: I1128 14:44:13.064184 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" podStartSLOduration=3.273711653 podStartE2EDuration="38.064166868s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.874038743 +0000 UTC m=+839.462017009" lastFinishedPulling="2025-11-28 14:44:11.664493958 +0000 UTC m=+874.252472224" observedRunningTime="2025-11-28 14:44:13.062872805 +0000 UTC m=+875.650851071" watchObservedRunningTime="2025-11-28 14:44:13.064166868 +0000 UTC m=+875.652145144" Nov 28 14:44:13 crc kubenswrapper[4817]: I1128 14:44:13.084958 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" podStartSLOduration=3.293895764 podStartE2EDuration="38.084936615s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:36.718344703 +0000 UTC m=+839.306322969" lastFinishedPulling="2025-11-28 14:44:11.509385554 +0000 UTC m=+874.097363820" observedRunningTime="2025-11-28 14:44:13.081209352 +0000 UTC m=+875.669187618" watchObservedRunningTime="2025-11-28 14:44:13.084936615 +0000 UTC m=+875.672914881" Nov 28 14:44:15 crc kubenswrapper[4817]: I1128 14:44:15.020158 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" event={"ID":"18412c56-76ce-48a3-81f2-3cbc4d090df5","Type":"ContainerStarted","Data":"8547ff09b329dc52c3f1391e97caa8410fcc320b6ab74b12893b5daa210a8924"} Nov 28 14:44:15 crc kubenswrapper[4817]: I1128 14:44:15.043438 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" podStartSLOduration=3.456479576 podStartE2EDuration="40.043412634s" podCreationTimestamp="2025-11-28 14:43:35 +0000 UTC" firstStartedPulling="2025-11-28 14:43:37.31705359 +0000 UTC m=+839.905031856" lastFinishedPulling="2025-11-28 14:44:13.903986648 +0000 UTC m=+876.491964914" observedRunningTime="2025-11-28 14:44:15.036071521 +0000 UTC m=+877.624049787" watchObservedRunningTime="2025-11-28 14:44:15.043412634 +0000 UTC m=+877.631390910" Nov 28 14:44:16 crc kubenswrapper[4817]: I1128 14:44:16.028250 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:44:16 crc kubenswrapper[4817]: I1128 14:44:16.147601 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-pmw8p" Nov 28 14:44:16 crc kubenswrapper[4817]: I1128 14:44:16.148342 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-k7tbk" Nov 28 14:44:16 crc kubenswrapper[4817]: I1128 14:44:16.164958 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-wzw2n" Nov 28 14:44:16 crc kubenswrapper[4817]: I1128 14:44:16.309365 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-8fqs4" Nov 28 14:44:16 crc kubenswrapper[4817]: I1128 14:44:16.362777 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5b7b8977cb-jkx7n" Nov 28 14:44:21 crc kubenswrapper[4817]: I1128 14:44:21.496863 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-w9qqs" Nov 28 14:44:21 crc kubenswrapper[4817]: I1128 14:44:21.883924 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm" Nov 28 14:44:25 crc kubenswrapper[4817]: I1128 14:44:25.759481 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-75m4b" Nov 28 14:44:25 crc kubenswrapper[4817]: I1128 14:44:25.810617 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-4fj4z" Nov 28 14:44:25 crc kubenswrapper[4817]: I1128 14:44:25.910142 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-vh4rl" Nov 28 14:44:25 crc kubenswrapper[4817]: I1128 14:44:25.932238 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4phnt" Nov 28 14:44:25 crc kubenswrapper[4817]: I1128 14:44:25.990643 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" Nov 28 14:44:26 crc kubenswrapper[4817]: I1128 14:44:26.183011 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-gv25c" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.889563 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4tbc5"] Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.899445 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.905118 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-789jc" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.906774 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4tbc5"] Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.912298 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.912672 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.912775 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.959783 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tktrb"] Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.961259 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.969241 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.972782 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48a72a9-32cd-46fa-93e6-3607caeba4cc-config\") pod \"dnsmasq-dns-675f4bcbfc-4tbc5\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.972849 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8zgp\" (UniqueName: \"kubernetes.io/projected/a48a72a9-32cd-46fa-93e6-3607caeba4cc-kube-api-access-q8zgp\") pod \"dnsmasq-dns-675f4bcbfc-4tbc5\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:41 crc kubenswrapper[4817]: I1128 14:44:41.990780 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tktrb"] Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.074141 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tq2h\" (UniqueName: \"kubernetes.io/projected/59ae3f67-fbf4-42c7-aa63-c7d802675afd-kube-api-access-2tq2h\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.074261 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.074299 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48a72a9-32cd-46fa-93e6-3607caeba4cc-config\") pod \"dnsmasq-dns-675f4bcbfc-4tbc5\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.074332 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-config\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.074368 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8zgp\" (UniqueName: \"kubernetes.io/projected/a48a72a9-32cd-46fa-93e6-3607caeba4cc-kube-api-access-q8zgp\") pod \"dnsmasq-dns-675f4bcbfc-4tbc5\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.075579 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48a72a9-32cd-46fa-93e6-3607caeba4cc-config\") pod \"dnsmasq-dns-675f4bcbfc-4tbc5\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.097897 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8zgp\" (UniqueName: \"kubernetes.io/projected/a48a72a9-32cd-46fa-93e6-3607caeba4cc-kube-api-access-q8zgp\") pod \"dnsmasq-dns-675f4bcbfc-4tbc5\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.175400 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tq2h\" (UniqueName: \"kubernetes.io/projected/59ae3f67-fbf4-42c7-aa63-c7d802675afd-kube-api-access-2tq2h\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.175492 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.175520 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-config\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.176242 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-config\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.177028 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.191683 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tq2h\" (UniqueName: \"kubernetes.io/projected/59ae3f67-fbf4-42c7-aa63-c7d802675afd-kube-api-access-2tq2h\") pod \"dnsmasq-dns-78dd6ddcc-tktrb\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.236687 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.287097 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.666325 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4tbc5"] Nov 28 14:44:42 crc kubenswrapper[4817]: I1128 14:44:42.746714 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tktrb"] Nov 28 14:44:42 crc kubenswrapper[4817]: W1128 14:44:42.749492 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59ae3f67_fbf4_42c7_aa63_c7d802675afd.slice/crio-41db9b251b3850caf471e9c5365c436fb714dce23822a5d303774f13a10a4b41 WatchSource:0}: Error finding container 41db9b251b3850caf471e9c5365c436fb714dce23822a5d303774f13a10a4b41: Status 404 returned error can't find the container with id 41db9b251b3850caf471e9c5365c436fb714dce23822a5d303774f13a10a4b41 Nov 28 14:44:43 crc kubenswrapper[4817]: I1128 14:44:43.316663 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" event={"ID":"a48a72a9-32cd-46fa-93e6-3607caeba4cc","Type":"ContainerStarted","Data":"314bc1a499bfa0a80e118b567e72cd69c1061f7b5ab3d3ca4a9285f7bb157544"} Nov 28 14:44:43 crc kubenswrapper[4817]: I1128 14:44:43.318084 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" event={"ID":"59ae3f67-fbf4-42c7-aa63-c7d802675afd","Type":"ContainerStarted","Data":"41db9b251b3850caf471e9c5365c436fb714dce23822a5d303774f13a10a4b41"} Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.000638 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4tbc5"] Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.019191 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p248n"] Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.021269 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.030112 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p248n"] Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.123016 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gmgf\" (UniqueName: \"kubernetes.io/projected/34bea336-efd1-4432-8ccf-afc87bee4381-kube-api-access-7gmgf\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.123464 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.123493 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-config\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.224226 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gmgf\" (UniqueName: \"kubernetes.io/projected/34bea336-efd1-4432-8ccf-afc87bee4381-kube-api-access-7gmgf\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.224282 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.224305 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-config\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.225266 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-config\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.225478 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-dns-svc\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.260250 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gmgf\" (UniqueName: \"kubernetes.io/projected/34bea336-efd1-4432-8ccf-afc87bee4381-kube-api-access-7gmgf\") pod \"dnsmasq-dns-666b6646f7-p248n\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.325797 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tktrb"] Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.350820 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.359591 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2gk4w"] Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.361234 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.381231 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2gk4w"] Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.428682 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-config\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.428740 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcv4v\" (UniqueName: \"kubernetes.io/projected/bd260782-cb7b-446c-8222-e28050effe0b-kube-api-access-zcv4v\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.428771 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.533526 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-config\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.533574 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcv4v\" (UniqueName: \"kubernetes.io/projected/bd260782-cb7b-446c-8222-e28050effe0b-kube-api-access-zcv4v\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.533607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.534473 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-config\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.536847 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.556597 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcv4v\" (UniqueName: \"kubernetes.io/projected/bd260782-cb7b-446c-8222-e28050effe0b-kube-api-access-zcv4v\") pod \"dnsmasq-dns-57d769cc4f-2gk4w\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.692573 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:44:45 crc kubenswrapper[4817]: I1128 14:44:45.951160 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p248n"] Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.186444 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2gk4w"] Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.196838 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.198068 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.204203 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.204487 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.205306 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.205454 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.205591 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.205830 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.216940 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6vf9j" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.228585 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256278 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256318 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256334 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63b71752-f06a-4d1a-9d80-85142bbc3610-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256349 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256420 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-server-conf\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256441 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgh2\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-kube-api-access-fdgh2\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256561 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63b71752-f06a-4d1a-9d80-85142bbc3610-pod-info\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256583 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256597 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256621 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-config-data\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.256644 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.357588 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63b71752-f06a-4d1a-9d80-85142bbc3610-pod-info\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359406 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359437 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359475 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-config-data\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359516 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359555 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359578 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359599 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63b71752-f06a-4d1a-9d80-85142bbc3610-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359623 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359653 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-server-conf\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.359678 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgh2\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-kube-api-access-fdgh2\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.360846 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-config-data\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.360993 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.361480 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.361763 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.362565 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.364483 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-server-conf\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.365445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.366067 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63b71752-f06a-4d1a-9d80-85142bbc3610-pod-info\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.366394 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63b71752-f06a-4d1a-9d80-85142bbc3610-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.370526 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.376765 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" event={"ID":"bd260782-cb7b-446c-8222-e28050effe0b","Type":"ContainerStarted","Data":"4f17be5e5bd9cf44bfa675c7cdeb089694b2b950cd0387e0a944b52df64132a8"} Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.382482 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p248n" event={"ID":"34bea336-efd1-4432-8ccf-afc87bee4381","Type":"ContainerStarted","Data":"c007ae040ebc42bf71e9916f77c3edd74935748098bfe580641b429284558e4e"} Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.384639 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgh2\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-kube-api-access-fdgh2\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.398985 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.489951 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.493608 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.496083 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.496118 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.496779 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.496990 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.498007 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qh99g" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.498049 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.498015 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.498900 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.528502 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667547 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667600 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667626 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667657 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d340d5c2-67ed-48da-826f-e1e4c244832b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667690 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2d84\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-kube-api-access-f2d84\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667762 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d340d5c2-67ed-48da-826f-e1e4c244832b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667802 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667830 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667850 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667870 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.667917 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769533 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d340d5c2-67ed-48da-826f-e1e4c244832b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769574 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769631 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769647 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769673 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769748 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769770 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769786 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769803 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d340d5c2-67ed-48da-826f-e1e4c244832b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769830 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2d84\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-kube-api-access-f2d84\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.769915 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.770097 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.770320 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.773822 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.774924 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.775435 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.776888 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d340d5c2-67ed-48da-826f-e1e4c244832b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.778166 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.779249 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.781975 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d340d5c2-67ed-48da-826f-e1e4c244832b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.785018 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2d84\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-kube-api-access-f2d84\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.822259 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:46 crc kubenswrapper[4817]: I1128 14:44:46.832083 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.075492 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:44:47 crc kubenswrapper[4817]: W1128 14:44:47.090602 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b71752_f06a_4d1a_9d80_85142bbc3610.slice/crio-185a4b26ff5692655011f475f0a810ea3638e19206e8f3a960240ea602406b6b WatchSource:0}: Error finding container 185a4b26ff5692655011f475f0a810ea3638e19206e8f3a960240ea602406b6b: Status 404 returned error can't find the container with id 185a4b26ff5692655011f475f0a810ea3638e19206e8f3a960240ea602406b6b Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.359794 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.394909 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63b71752-f06a-4d1a-9d80-85142bbc3610","Type":"ContainerStarted","Data":"185a4b26ff5692655011f475f0a810ea3638e19206e8f3a960240ea602406b6b"} Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.735093 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.737096 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.752083 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.753282 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.753410 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4zn4k" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.753535 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.753623 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.759087 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.886900 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-config-data-default\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.886940 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.886957 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-kolla-config\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.886982 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd723f67-725a-4942-87f2-c646eddcd172-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.887010 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdk8d\" (UniqueName: \"kubernetes.io/projected/bd723f67-725a-4942-87f2-c646eddcd172-kube-api-access-kdk8d\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.887036 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.887132 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bd723f67-725a-4942-87f2-c646eddcd172-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.887226 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd723f67-725a-4942-87f2-c646eddcd172-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990139 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-config-data-default\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.989030 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-config-data-default\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990259 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990299 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-kolla-config\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990339 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd723f67-725a-4942-87f2-c646eddcd172-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990388 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdk8d\" (UniqueName: \"kubernetes.io/projected/bd723f67-725a-4942-87f2-c646eddcd172-kube-api-access-kdk8d\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990422 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990469 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bd723f67-725a-4942-87f2-c646eddcd172-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.990533 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd723f67-725a-4942-87f2-c646eddcd172-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.991586 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.991683 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bd723f67-725a-4942-87f2-c646eddcd172-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.992856 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.993585 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bd723f67-725a-4942-87f2-c646eddcd172-kolla-config\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.997099 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd723f67-725a-4942-87f2-c646eddcd172-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:47 crc kubenswrapper[4817]: I1128 14:44:47.997206 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd723f67-725a-4942-87f2-c646eddcd172-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:48 crc kubenswrapper[4817]: I1128 14:44:48.016298 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdk8d\" (UniqueName: \"kubernetes.io/projected/bd723f67-725a-4942-87f2-c646eddcd172-kube-api-access-kdk8d\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:48 crc kubenswrapper[4817]: I1128 14:44:48.016945 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"bd723f67-725a-4942-87f2-c646eddcd172\") " pod="openstack/openstack-galera-0" Nov 28 14:44:48 crc kubenswrapper[4817]: I1128 14:44:48.069938 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.308042 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.309426 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.314480 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.314753 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-h7gdp" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.314843 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.314936 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.327348 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.422454 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.422515 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.422554 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.423548 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.423624 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.423671 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.423698 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mpzc\" (UniqueName: \"kubernetes.io/projected/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-kube-api-access-5mpzc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.423817 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527540 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527615 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527636 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527666 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mpzc\" (UniqueName: \"kubernetes.io/projected/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-kube-api-access-5mpzc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527691 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527778 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527802 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.527840 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.528312 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.530962 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.531506 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.532526 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.536794 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.538310 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.542490 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.558200 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mpzc\" (UniqueName: \"kubernetes.io/projected/ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d-kube-api-access-5mpzc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.596511 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d\") " pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.641352 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.718915 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.720159 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.723117 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.723395 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-6tkfh" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.723539 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.729638 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.832477 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789d2be2-8615-428f-bc54-68d019823a99-combined-ca-bundle\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.832543 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/789d2be2-8615-428f-bc54-68d019823a99-kolla-config\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.832577 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwmp\" (UniqueName: \"kubernetes.io/projected/789d2be2-8615-428f-bc54-68d019823a99-kube-api-access-mgwmp\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.832642 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/789d2be2-8615-428f-bc54-68d019823a99-memcached-tls-certs\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.832661 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/789d2be2-8615-428f-bc54-68d019823a99-config-data\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.933435 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/789d2be2-8615-428f-bc54-68d019823a99-kolla-config\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.933537 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwmp\" (UniqueName: \"kubernetes.io/projected/789d2be2-8615-428f-bc54-68d019823a99-kube-api-access-mgwmp\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.933603 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/789d2be2-8615-428f-bc54-68d019823a99-memcached-tls-certs\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.933632 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/789d2be2-8615-428f-bc54-68d019823a99-config-data\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.933656 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789d2be2-8615-428f-bc54-68d019823a99-combined-ca-bundle\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.934349 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/789d2be2-8615-428f-bc54-68d019823a99-kolla-config\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.943002 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/789d2be2-8615-428f-bc54-68d019823a99-config-data\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.952376 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwmp\" (UniqueName: \"kubernetes.io/projected/789d2be2-8615-428f-bc54-68d019823a99-kube-api-access-mgwmp\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.961767 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789d2be2-8615-428f-bc54-68d019823a99-combined-ca-bundle\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:49 crc kubenswrapper[4817]: I1128 14:44:49.968451 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/789d2be2-8615-428f-bc54-68d019823a99-memcached-tls-certs\") pod \"memcached-0\" (UID: \"789d2be2-8615-428f-bc54-68d019823a99\") " pod="openstack/memcached-0" Nov 28 14:44:50 crc kubenswrapper[4817]: I1128 14:44:50.046427 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.760695 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.761885 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.764575 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-2jsbh" Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.767037 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.862169 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7jxl\" (UniqueName: \"kubernetes.io/projected/bef81fb8-da0d-4735-8882-5fc41530730c-kube-api-access-m7jxl\") pod \"kube-state-metrics-0\" (UID: \"bef81fb8-da0d-4735-8882-5fc41530730c\") " pod="openstack/kube-state-metrics-0" Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.963932 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7jxl\" (UniqueName: \"kubernetes.io/projected/bef81fb8-da0d-4735-8882-5fc41530730c-kube-api-access-m7jxl\") pod \"kube-state-metrics-0\" (UID: \"bef81fb8-da0d-4735-8882-5fc41530730c\") " pod="openstack/kube-state-metrics-0" Nov 28 14:44:51 crc kubenswrapper[4817]: I1128 14:44:51.982978 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7jxl\" (UniqueName: \"kubernetes.io/projected/bef81fb8-da0d-4735-8882-5fc41530730c-kube-api-access-m7jxl\") pod \"kube-state-metrics-0\" (UID: \"bef81fb8-da0d-4735-8882-5fc41530730c\") " pod="openstack/kube-state-metrics-0" Nov 28 14:44:52 crc kubenswrapper[4817]: I1128 14:44:52.086383 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.339410 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pwgz2"] Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.341151 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.343190 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.343338 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.344643 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-pbg5b" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.351299 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pwgz2"] Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.416864 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-vz6mt"] Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.418372 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.432703 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vz6mt"] Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515375 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng4qs\" (UniqueName: \"kubernetes.io/projected/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-kube-api-access-ng4qs\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515435 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-scripts\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515469 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-ovn-controller-tls-certs\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515739 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-combined-ca-bundle\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515770 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-run-ovn\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515793 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-run\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.515810 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-log-ovn\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.616853 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-run\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.616904 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-log-ovn\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.616937 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng4qs\" (UniqueName: \"kubernetes.io/projected/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-kube-api-access-ng4qs\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.616982 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-scripts\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617016 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-lib\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617036 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-etc-ovs\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617056 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55qj7\" (UniqueName: \"kubernetes.io/projected/40df6e7a-e727-4386-bca8-aeff35121dbb-kube-api-access-55qj7\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617127 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-ovn-controller-tls-certs\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617190 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40df6e7a-e727-4386-bca8-aeff35121dbb-scripts\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617221 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-run\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617287 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-log\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617372 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-combined-ca-bundle\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617437 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-run-ovn\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617609 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-log-ovn\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.617677 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-run-ovn\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.618323 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-var-run\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.619321 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-scripts\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.622315 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-ovn-controller-tls-certs\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.627187 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-combined-ca-bundle\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.635815 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng4qs\" (UniqueName: \"kubernetes.io/projected/0f58e7f4-f831-4f54-91b6-a5e97bf5152c-kube-api-access-ng4qs\") pod \"ovn-controller-pwgz2\" (UID: \"0f58e7f4-f831-4f54-91b6-a5e97bf5152c\") " pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.660790 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.690984 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.692256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.693918 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.694393 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.694999 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rlftc" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.696477 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.701816 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.703519 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.718855 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-lib\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.718893 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-etc-ovs\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.718911 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55qj7\" (UniqueName: \"kubernetes.io/projected/40df6e7a-e727-4386-bca8-aeff35121dbb-kube-api-access-55qj7\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.718937 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40df6e7a-e727-4386-bca8-aeff35121dbb-scripts\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.718960 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-run\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.718981 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-log\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.719323 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-log\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.719471 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-lib\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.719606 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-etc-ovs\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.721483 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40df6e7a-e727-4386-bca8-aeff35121dbb-var-run\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.722160 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40df6e7a-e727-4386-bca8-aeff35121dbb-scripts\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.740879 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55qj7\" (UniqueName: \"kubernetes.io/projected/40df6e7a-e727-4386-bca8-aeff35121dbb-kube-api-access-55qj7\") pod \"ovn-controller-ovs-vz6mt\" (UID: \"40df6e7a-e727-4386-bca8-aeff35121dbb\") " pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.819972 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820197 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4zcm\" (UniqueName: \"kubernetes.io/projected/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-kube-api-access-x4zcm\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820351 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820440 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820508 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820597 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820745 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.820991 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-config\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922401 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922455 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922525 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922554 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-config\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922584 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922615 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4zcm\" (UniqueName: \"kubernetes.io/projected/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-kube-api-access-x4zcm\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922664 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.922681 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.923228 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.923930 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.923948 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-config\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.926123 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.926820 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.926911 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.927207 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.952086 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4zcm\" (UniqueName: \"kubernetes.io/projected/e3b52366-bcd8-4685-9f36-2adaea4e6cc4-kube-api-access-x4zcm\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:55 crc kubenswrapper[4817]: I1128 14:44:55.964204 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e3b52366-bcd8-4685-9f36-2adaea4e6cc4\") " pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:56 crc kubenswrapper[4817]: I1128 14:44:56.030275 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 14:44:56 crc kubenswrapper[4817]: I1128 14:44:56.034322 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:44:57 crc kubenswrapper[4817]: I1128 14:44:57.475545 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d340d5c2-67ed-48da-826f-e1e4c244832b","Type":"ContainerStarted","Data":"a4c4f9342056520de6653dffa97f90343f2042f50c507143a12861b13ea6b4bf"} Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.279145 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.281256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.284941 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.289174 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.289366 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.292093 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-gq2c8" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.322063 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418584 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418641 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418694 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418753 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ff9385f1-3c29-48be-bc80-806380c08646-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418824 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff9385f1-3c29-48be-bc80-806380c08646-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418857 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418889 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9385f1-3c29-48be-bc80-806380c08646-config\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.418958 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkmm5\" (UniqueName: \"kubernetes.io/projected/ff9385f1-3c29-48be-bc80-806380c08646-kube-api-access-vkmm5\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523419 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523535 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523594 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ff9385f1-3c29-48be-bc80-806380c08646-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523667 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff9385f1-3c29-48be-bc80-806380c08646-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523713 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523877 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9385f1-3c29-48be-bc80-806380c08646-config\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.523971 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkmm5\" (UniqueName: \"kubernetes.io/projected/ff9385f1-3c29-48be-bc80-806380c08646-kube-api-access-vkmm5\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.524074 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.525572 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff9385f1-3c29-48be-bc80-806380c08646-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.525819 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ff9385f1-3c29-48be-bc80-806380c08646-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.525848 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.526519 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9385f1-3c29-48be-bc80-806380c08646-config\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.530939 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.539769 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.543416 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9385f1-3c29-48be-bc80-806380c08646-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.553472 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkmm5\" (UniqueName: \"kubernetes.io/projected/ff9385f1-3c29-48be-bc80-806380c08646-kube-api-access-vkmm5\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.563061 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ff9385f1-3c29-48be-bc80-806380c08646\") " pod="openstack/ovsdbserver-sb-0" Nov 28 14:44:59 crc kubenswrapper[4817]: I1128 14:44:59.614178 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.162162 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd"] Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.163775 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.169503 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.169569 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.186891 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd"] Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.338755 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75e59a9f-fb61-4364-ace0-8940f22583cb-secret-volume\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.338820 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xhmm\" (UniqueName: \"kubernetes.io/projected/75e59a9f-fb61-4364-ace0-8940f22583cb-kube-api-access-9xhmm\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.338915 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75e59a9f-fb61-4364-ace0-8940f22583cb-config-volume\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.440340 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75e59a9f-fb61-4364-ace0-8940f22583cb-config-volume\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.440475 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75e59a9f-fb61-4364-ace0-8940f22583cb-secret-volume\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.440513 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xhmm\" (UniqueName: \"kubernetes.io/projected/75e59a9f-fb61-4364-ace0-8940f22583cb-kube-api-access-9xhmm\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.441424 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75e59a9f-fb61-4364-ace0-8940f22583cb-config-volume\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.445164 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75e59a9f-fb61-4364-ace0-8940f22583cb-secret-volume\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.462579 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xhmm\" (UniqueName: \"kubernetes.io/projected/75e59a9f-fb61-4364-ace0-8940f22583cb-kube-api-access-9xhmm\") pod \"collect-profiles-29405685-drswd\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:00 crc kubenswrapper[4817]: I1128 14:45:00.559708 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:06 crc kubenswrapper[4817]: I1128 14:45:06.775317 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:45:12 crc kubenswrapper[4817]: I1128 14:45:12.822896 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-blkhg"] Nov 28 14:45:12 crc kubenswrapper[4817]: I1128 14:45:12.825503 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:12 crc kubenswrapper[4817]: I1128 14:45:12.836982 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-blkhg"] Nov 28 14:45:12 crc kubenswrapper[4817]: I1128 14:45:12.969142 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-utilities\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:12 crc kubenswrapper[4817]: I1128 14:45:12.969191 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w28r6\" (UniqueName: \"kubernetes.io/projected/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-kube-api-access-w28r6\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:12 crc kubenswrapper[4817]: I1128 14:45:12.969274 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-catalog-content\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.070413 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-utilities\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.070464 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w28r6\" (UniqueName: \"kubernetes.io/projected/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-kube-api-access-w28r6\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.070540 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-catalog-content\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.070993 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-utilities\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.071012 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-catalog-content\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.100922 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w28r6\" (UniqueName: \"kubernetes.io/projected/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-kube-api-access-w28r6\") pod \"community-operators-blkhg\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.156035 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:13 crc kubenswrapper[4817]: I1128 14:45:13.607646 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bef81fb8-da0d-4735-8882-5fc41530730c","Type":"ContainerStarted","Data":"8aa73995084e5522af7efdd0429259d5167e3bc809b9c8d546228ced98debf08"} Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.224582 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.224847 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zcv4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-2gk4w_openstack(bd260782-cb7b-446c-8222-e28050effe0b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.225980 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" podUID="bd260782-cb7b-446c-8222-e28050effe0b" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.233711 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.233947 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q8zgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-4tbc5_openstack(a48a72a9-32cd-46fa-93e6-3607caeba4cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.235071 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" podUID="a48a72a9-32cd-46fa-93e6-3607caeba4cc" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.250251 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.250339 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.250422 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2tq2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-tktrb_openstack(59ae3f67-fbf4-42c7-aa63-c7d802675afd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.250598 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7gmgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-p248n_openstack(34bea336-efd1-4432-8ccf-afc87bee4381): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.251784 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" podUID="59ae3f67-fbf4-42c7-aa63-c7d802675afd" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.251855 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-p248n" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.619058 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" podUID="bd260782-cb7b-446c-8222-e28050effe0b" Nov 28 14:45:14 crc kubenswrapper[4817]: E1128 14:45:14.620080 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-p248n" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" Nov 28 14:45:14 crc kubenswrapper[4817]: I1128 14:45:14.717034 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 14:45:14 crc kubenswrapper[4817]: W1128 14:45:14.743855 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd723f67_725a_4942_87f2_c646eddcd172.slice/crio-c9cc434ef172d327d2f501b990b1cab211d66897b4cd4d4197535d2c1586f857 WatchSource:0}: Error finding container c9cc434ef172d327d2f501b990b1cab211d66897b4cd4d4197535d2c1586f857: Status 404 returned error can't find the container with id c9cc434ef172d327d2f501b990b1cab211d66897b4cd4d4197535d2c1586f857 Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.162908 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pwgz2"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.173391 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-blkhg"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.259697 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.301917 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.311702 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.327671 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.404588 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 14:45:15 crc kubenswrapper[4817]: W1128 14:45:15.416456 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod789d2be2_8615_428f_bc54_68d019823a99.slice/crio-c66d4d935ba59cf13f748a4ae0b17d18e29fa5f214d3ac0bdd2ce4ab390040b0 WatchSource:0}: Error finding container c66d4d935ba59cf13f748a4ae0b17d18e29fa5f214d3ac0bdd2ce4ab390040b0: Status 404 returned error can't find the container with id c66d4d935ba59cf13f748a4ae0b17d18e29fa5f214d3ac0bdd2ce4ab390040b0 Nov 28 14:45:15 crc kubenswrapper[4817]: W1128 14:45:15.417296 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba08a213_ad14_4a0e_a7e5_c09ffa3cde3d.slice/crio-f0be3f33fbd022be8a0347350abfd3f5be7d50585bcbabcf94f94eec90b0a1f7 WatchSource:0}: Error finding container f0be3f33fbd022be8a0347350abfd3f5be7d50585bcbabcf94f94eec90b0a1f7: Status 404 returned error can't find the container with id f0be3f33fbd022be8a0347350abfd3f5be7d50585bcbabcf94f94eec90b0a1f7 Nov 28 14:45:15 crc kubenswrapper[4817]: W1128 14:45:15.424797 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75e59a9f_fb61_4364_ace0_8940f22583cb.slice/crio-8168ed6e1dcf20d35c2df4fa9d631902bf413772de152149847c555ae261e267 WatchSource:0}: Error finding container 8168ed6e1dcf20d35c2df4fa9d631902bf413772de152149847c555ae261e267: Status 404 returned error can't find the container with id 8168ed6e1dcf20d35c2df4fa9d631902bf413772de152149847c555ae261e267 Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.497651 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vz6mt"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.568394 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.575497 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.624669 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blkhg" event={"ID":"c3d3ce03-7abf-4cd7-8394-192f8c56ba51","Type":"ContainerStarted","Data":"2849bcba6b796f8c8c25ecbacbc185fc1005c347d1875ef33d76c83b98e53cbb"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.626116 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" event={"ID":"59ae3f67-fbf4-42c7-aa63-c7d802675afd","Type":"ContainerDied","Data":"41db9b251b3850caf471e9c5365c436fb714dce23822a5d303774f13a10a4b41"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.626125 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tktrb" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.627831 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" event={"ID":"a48a72a9-32cd-46fa-93e6-3607caeba4cc","Type":"ContainerDied","Data":"314bc1a499bfa0a80e118b567e72cd69c1061f7b5ab3d3ca4a9285f7bb157544"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.627883 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4tbc5" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.629910 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"789d2be2-8615-428f-bc54-68d019823a99","Type":"ContainerStarted","Data":"c66d4d935ba59cf13f748a4ae0b17d18e29fa5f214d3ac0bdd2ce4ab390040b0"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.631629 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d","Type":"ContainerStarted","Data":"f0be3f33fbd022be8a0347350abfd3f5be7d50585bcbabcf94f94eec90b0a1f7"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.633079 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" event={"ID":"75e59a9f-fb61-4364-ace0-8940f22583cb","Type":"ContainerStarted","Data":"8168ed6e1dcf20d35c2df4fa9d631902bf413772de152149847c555ae261e267"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.635661 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2" event={"ID":"0f58e7f4-f831-4f54-91b6-a5e97bf5152c","Type":"ContainerStarted","Data":"3d53729186bbbf2559ef83cf066f1ec8187694f86d098833a228cf62f43b858f"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.636859 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bd723f67-725a-4942-87f2-c646eddcd172","Type":"ContainerStarted","Data":"c9cc434ef172d327d2f501b990b1cab211d66897b4cd4d4197535d2c1586f857"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.637897 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e3b52366-bcd8-4685-9f36-2adaea4e6cc4","Type":"ContainerStarted","Data":"421e3ceeef30a92dc6ad6c6abc8d1668de9c718e24a456f41c0e1ec677a6c5cd"} Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.717994 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-config\") pod \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.718089 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tq2h\" (UniqueName: \"kubernetes.io/projected/59ae3f67-fbf4-42c7-aa63-c7d802675afd-kube-api-access-2tq2h\") pod \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.718184 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8zgp\" (UniqueName: \"kubernetes.io/projected/a48a72a9-32cd-46fa-93e6-3607caeba4cc-kube-api-access-q8zgp\") pod \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.718251 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48a72a9-32cd-46fa-93e6-3607caeba4cc-config\") pod \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\" (UID: \"a48a72a9-32cd-46fa-93e6-3607caeba4cc\") " Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.718357 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-dns-svc\") pod \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\" (UID: \"59ae3f67-fbf4-42c7-aa63-c7d802675afd\") " Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.718925 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-config" (OuterVolumeSpecName: "config") pod "59ae3f67-fbf4-42c7-aa63-c7d802675afd" (UID: "59ae3f67-fbf4-42c7-aa63-c7d802675afd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.719672 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a48a72a9-32cd-46fa-93e6-3607caeba4cc-config" (OuterVolumeSpecName: "config") pod "a48a72a9-32cd-46fa-93e6-3607caeba4cc" (UID: "a48a72a9-32cd-46fa-93e6-3607caeba4cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.721115 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59ae3f67-fbf4-42c7-aa63-c7d802675afd" (UID: "59ae3f67-fbf4-42c7-aa63-c7d802675afd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.727189 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59ae3f67-fbf4-42c7-aa63-c7d802675afd-kube-api-access-2tq2h" (OuterVolumeSpecName: "kube-api-access-2tq2h") pod "59ae3f67-fbf4-42c7-aa63-c7d802675afd" (UID: "59ae3f67-fbf4-42c7-aa63-c7d802675afd"). InnerVolumeSpecName "kube-api-access-2tq2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.729766 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a48a72a9-32cd-46fa-93e6-3607caeba4cc-kube-api-access-q8zgp" (OuterVolumeSpecName: "kube-api-access-q8zgp") pod "a48a72a9-32cd-46fa-93e6-3607caeba4cc" (UID: "a48a72a9-32cd-46fa-93e6-3607caeba4cc"). InnerVolumeSpecName "kube-api-access-q8zgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.821150 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.821614 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tq2h\" (UniqueName: \"kubernetes.io/projected/59ae3f67-fbf4-42c7-aa63-c7d802675afd-kube-api-access-2tq2h\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.821654 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8zgp\" (UniqueName: \"kubernetes.io/projected/a48a72a9-32cd-46fa-93e6-3607caeba4cc-kube-api-access-q8zgp\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.821668 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48a72a9-32cd-46fa-93e6-3607caeba4cc-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.821681 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59ae3f67-fbf4-42c7-aa63-c7d802675afd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.976766 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tktrb"] Nov 28 14:45:15 crc kubenswrapper[4817]: I1128 14:45:15.992727 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tktrb"] Nov 28 14:45:16 crc kubenswrapper[4817]: I1128 14:45:16.008527 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4tbc5"] Nov 28 14:45:16 crc kubenswrapper[4817]: I1128 14:45:16.015747 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4tbc5"] Nov 28 14:45:16 crc kubenswrapper[4817]: I1128 14:45:16.645477 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ff9385f1-3c29-48be-bc80-806380c08646","Type":"ContainerStarted","Data":"850653a5bf802b6b9d3b55d0948a1ab5f23469c405c052721bbd754b7eb0ce70"} Nov 28 14:45:16 crc kubenswrapper[4817]: I1128 14:45:16.647219 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vz6mt" event={"ID":"40df6e7a-e727-4386-bca8-aeff35121dbb","Type":"ContainerStarted","Data":"b577eaad9865a776b2912e9a43c718678da0a30e4c5586a05fc208c50805f40c"} Nov 28 14:45:16 crc kubenswrapper[4817]: I1128 14:45:16.650150 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63b71752-f06a-4d1a-9d80-85142bbc3610","Type":"ContainerStarted","Data":"290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4"} Nov 28 14:45:17 crc kubenswrapper[4817]: I1128 14:45:17.663210 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d340d5c2-67ed-48da-826f-e1e4c244832b","Type":"ContainerStarted","Data":"8ccf30fb29efa6e931504451a614334aecdfe48ab9668d3b04ee44037bd78dc6"} Nov 28 14:45:17 crc kubenswrapper[4817]: I1128 14:45:17.748021 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59ae3f67-fbf4-42c7-aa63-c7d802675afd" path="/var/lib/kubelet/pods/59ae3f67-fbf4-42c7-aa63-c7d802675afd/volumes" Nov 28 14:45:17 crc kubenswrapper[4817]: I1128 14:45:17.748659 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a48a72a9-32cd-46fa-93e6-3607caeba4cc" path="/var/lib/kubelet/pods/a48a72a9-32cd-46fa-93e6-3607caeba4cc/volumes" Nov 28 14:45:20 crc kubenswrapper[4817]: I1128 14:45:20.692329 4817 generic.go:334] "Generic (PLEG): container finished" podID="75e59a9f-fb61-4364-ace0-8940f22583cb" containerID="301e989706465d265ad00f51473a5f18b7d012a3e8f35245c2c2ac58a32f27db" exitCode=0 Nov 28 14:45:20 crc kubenswrapper[4817]: I1128 14:45:20.692414 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" event={"ID":"75e59a9f-fb61-4364-ace0-8940f22583cb","Type":"ContainerDied","Data":"301e989706465d265ad00f51473a5f18b7d012a3e8f35245c2c2ac58a32f27db"} Nov 28 14:45:20 crc kubenswrapper[4817]: I1128 14:45:20.696402 4817 generic.go:334] "Generic (PLEG): container finished" podID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerID="f704860adb9476fc82697e701a70dbc79a26b7b31cd0d05f9241a86e359d8a03" exitCode=0 Nov 28 14:45:20 crc kubenswrapper[4817]: I1128 14:45:20.696439 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blkhg" event={"ID":"c3d3ce03-7abf-4cd7-8394-192f8c56ba51","Type":"ContainerDied","Data":"f704860adb9476fc82697e701a70dbc79a26b7b31cd0d05f9241a86e359d8a03"} Nov 28 14:45:22 crc kubenswrapper[4817]: I1128 14:45:22.926478 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.028804 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75e59a9f-fb61-4364-ace0-8940f22583cb-secret-volume\") pod \"75e59a9f-fb61-4364-ace0-8940f22583cb\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.028897 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xhmm\" (UniqueName: \"kubernetes.io/projected/75e59a9f-fb61-4364-ace0-8940f22583cb-kube-api-access-9xhmm\") pod \"75e59a9f-fb61-4364-ace0-8940f22583cb\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.028937 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75e59a9f-fb61-4364-ace0-8940f22583cb-config-volume\") pod \"75e59a9f-fb61-4364-ace0-8940f22583cb\" (UID: \"75e59a9f-fb61-4364-ace0-8940f22583cb\") " Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.030372 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75e59a9f-fb61-4364-ace0-8940f22583cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "75e59a9f-fb61-4364-ace0-8940f22583cb" (UID: "75e59a9f-fb61-4364-ace0-8940f22583cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.038796 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e59a9f-fb61-4364-ace0-8940f22583cb-kube-api-access-9xhmm" (OuterVolumeSpecName: "kube-api-access-9xhmm") pod "75e59a9f-fb61-4364-ace0-8940f22583cb" (UID: "75e59a9f-fb61-4364-ace0-8940f22583cb"). InnerVolumeSpecName "kube-api-access-9xhmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.044253 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e59a9f-fb61-4364-ace0-8940f22583cb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "75e59a9f-fb61-4364-ace0-8940f22583cb" (UID: "75e59a9f-fb61-4364-ace0-8940f22583cb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.135597 4817 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75e59a9f-fb61-4364-ace0-8940f22583cb-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.135641 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xhmm\" (UniqueName: \"kubernetes.io/projected/75e59a9f-fb61-4364-ace0-8940f22583cb-kube-api-access-9xhmm\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.135659 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75e59a9f-fb61-4364-ace0-8940f22583cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.721424 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.721424 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd" event={"ID":"75e59a9f-fb61-4364-ace0-8940f22583cb","Type":"ContainerDied","Data":"8168ed6e1dcf20d35c2df4fa9d631902bf413772de152149847c555ae261e267"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.722019 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8168ed6e1dcf20d35c2df4fa9d631902bf413772de152149847c555ae261e267" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.724864 4817 generic.go:334] "Generic (PLEG): container finished" podID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerID="883a32419ad50c9bc2eb8dbe28886a53d9e7db848731eb8d66cfca1f1019eefb" exitCode=0 Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.725175 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blkhg" event={"ID":"c3d3ce03-7abf-4cd7-8394-192f8c56ba51","Type":"ContainerDied","Data":"883a32419ad50c9bc2eb8dbe28886a53d9e7db848731eb8d66cfca1f1019eefb"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.726851 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bd723f67-725a-4942-87f2-c646eddcd172","Type":"ContainerStarted","Data":"9c75bcc23dbb7e2a619a7f66c20bc7da173287c02298c626f12505394fe53fa5"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.730808 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ff9385f1-3c29-48be-bc80-806380c08646","Type":"ContainerStarted","Data":"ae168c422c711f0bb18e93d396efcc0d757ae26411d122f82162d248ef778c6b"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.735036 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bef81fb8-da0d-4735-8882-5fc41530730c","Type":"ContainerStarted","Data":"7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.736145 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.759079 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-pwgz2" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.759130 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e3b52366-bcd8-4685-9f36-2adaea4e6cc4","Type":"ContainerStarted","Data":"c19f88bcd8315ae900b5a44a454d476a4122149c97853a2778c642ac29e9c6b3"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.759163 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2" event={"ID":"0f58e7f4-f831-4f54-91b6-a5e97bf5152c","Type":"ContainerStarted","Data":"f73f2d658fb2e4d5d311ace58cacabf6698322b01f8baa225cf4d7dc909ff396"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.763546 4817 generic.go:334] "Generic (PLEG): container finished" podID="40df6e7a-e727-4386-bca8-aeff35121dbb" containerID="56dd5eddfa95902b55a7c7c7ce4670fb3e69cd99eacf57f080b6fe13fb376c9a" exitCode=0 Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.763655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vz6mt" event={"ID":"40df6e7a-e727-4386-bca8-aeff35121dbb","Type":"ContainerDied","Data":"56dd5eddfa95902b55a7c7c7ce4670fb3e69cd99eacf57f080b6fe13fb376c9a"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.767400 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"789d2be2-8615-428f-bc54-68d019823a99","Type":"ContainerStarted","Data":"5538cc809c584047a981ec22a3d58a06443321e4ee7be82a89577ee751288c3a"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.768157 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.776207 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d","Type":"ContainerStarted","Data":"d4c8acb4f4ca076ef548ee41a75a7e89b594e0bd8220ccd9a8bc20e38b254cf2"} Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.784559 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=23.51165718 podStartE2EDuration="32.784536187s" podCreationTimestamp="2025-11-28 14:44:51 +0000 UTC" firstStartedPulling="2025-11-28 14:45:13.507986546 +0000 UTC m=+936.095964822" lastFinishedPulling="2025-11-28 14:45:22.780865563 +0000 UTC m=+945.368843829" observedRunningTime="2025-11-28 14:45:23.780342993 +0000 UTC m=+946.368321269" watchObservedRunningTime="2025-11-28 14:45:23.784536187 +0000 UTC m=+946.372514463" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.822708 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pwgz2" podStartSLOduration=21.430694416 podStartE2EDuration="28.822680737s" podCreationTimestamp="2025-11-28 14:44:55 +0000 UTC" firstStartedPulling="2025-11-28 14:45:15.401351363 +0000 UTC m=+937.989329619" lastFinishedPulling="2025-11-28 14:45:22.793337674 +0000 UTC m=+945.381315940" observedRunningTime="2025-11-28 14:45:23.814049862 +0000 UTC m=+946.402028128" watchObservedRunningTime="2025-11-28 14:45:23.822680737 +0000 UTC m=+946.410659013" Nov 28 14:45:23 crc kubenswrapper[4817]: I1128 14:45:23.887616 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=27.55849176 podStartE2EDuration="34.887600755s" podCreationTimestamp="2025-11-28 14:44:49 +0000 UTC" firstStartedPulling="2025-11-28 14:45:15.419020043 +0000 UTC m=+938.006998309" lastFinishedPulling="2025-11-28 14:45:22.748129038 +0000 UTC m=+945.336107304" observedRunningTime="2025-11-28 14:45:23.864564651 +0000 UTC m=+946.452542917" watchObservedRunningTime="2025-11-28 14:45:23.887600755 +0000 UTC m=+946.475579021" Nov 28 14:45:24 crc kubenswrapper[4817]: I1128 14:45:24.798410 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vz6mt" event={"ID":"40df6e7a-e727-4386-bca8-aeff35121dbb","Type":"ContainerStarted","Data":"b1adcb7ccec12f5e820ebe92b2f6d5e582cb54b278de70487d1e901b84684ff7"} Nov 28 14:45:24 crc kubenswrapper[4817]: I1128 14:45:24.799191 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vz6mt" event={"ID":"40df6e7a-e727-4386-bca8-aeff35121dbb","Type":"ContainerStarted","Data":"2f035db1dc34c265d6befff9d7c97bca280b701f2efa4d5eecd59eace79e8b02"} Nov 28 14:45:24 crc kubenswrapper[4817]: I1128 14:45:24.799215 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:45:24 crc kubenswrapper[4817]: I1128 14:45:24.799250 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:45:25 crc kubenswrapper[4817]: I1128 14:45:25.815556 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blkhg" event={"ID":"c3d3ce03-7abf-4cd7-8394-192f8c56ba51","Type":"ContainerStarted","Data":"e0f32889994eb0364eca54cf0dd12567a88917189596c669f745f4fc455c1611"} Nov 28 14:45:25 crc kubenswrapper[4817]: I1128 14:45:25.844024 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-blkhg" podStartSLOduration=10.673764814 podStartE2EDuration="13.843988162s" podCreationTimestamp="2025-11-28 14:45:12 +0000 UTC" firstStartedPulling="2025-11-28 14:45:21.474073328 +0000 UTC m=+944.062051594" lastFinishedPulling="2025-11-28 14:45:24.644296676 +0000 UTC m=+947.232274942" observedRunningTime="2025-11-28 14:45:25.840027283 +0000 UTC m=+948.428005569" watchObservedRunningTime="2025-11-28 14:45:25.843988162 +0000 UTC m=+948.431966428" Nov 28 14:45:25 crc kubenswrapper[4817]: I1128 14:45:25.844354 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-vz6mt" podStartSLOduration=23.803794085 podStartE2EDuration="30.844348191s" podCreationTimestamp="2025-11-28 14:44:55 +0000 UTC" firstStartedPulling="2025-11-28 14:45:15.740511472 +0000 UTC m=+938.328489738" lastFinishedPulling="2025-11-28 14:45:22.781065578 +0000 UTC m=+945.369043844" observedRunningTime="2025-11-28 14:45:24.824030463 +0000 UTC m=+947.412008739" watchObservedRunningTime="2025-11-28 14:45:25.844348191 +0000 UTC m=+948.432326467" Nov 28 14:45:27 crc kubenswrapper[4817]: I1128 14:45:27.834040 4817 generic.go:334] "Generic (PLEG): container finished" podID="34bea336-efd1-4432-8ccf-afc87bee4381" containerID="2e28bf49af60e63d337794899a865b6df304e2e1d53f2d5fea9aa1a058ab7e7d" exitCode=0 Nov 28 14:45:27 crc kubenswrapper[4817]: I1128 14:45:27.834419 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p248n" event={"ID":"34bea336-efd1-4432-8ccf-afc87bee4381","Type":"ContainerDied","Data":"2e28bf49af60e63d337794899a865b6df304e2e1d53f2d5fea9aa1a058ab7e7d"} Nov 28 14:45:27 crc kubenswrapper[4817]: I1128 14:45:27.837676 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e3b52366-bcd8-4685-9f36-2adaea4e6cc4","Type":"ContainerStarted","Data":"5a7efee8c866b6c0a562ca810bf5bfc2249c91b2160543f8ba263737569b6bfb"} Nov 28 14:45:27 crc kubenswrapper[4817]: I1128 14:45:27.841117 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ff9385f1-3c29-48be-bc80-806380c08646","Type":"ContainerStarted","Data":"134670ddbdcd6e8373f7e82e7ad0b49ab48f1092c8dd799ed7626442f9911b7d"} Nov 28 14:45:27 crc kubenswrapper[4817]: I1128 14:45:27.890330 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=21.895577046 podStartE2EDuration="33.890295751s" podCreationTimestamp="2025-11-28 14:44:54 +0000 UTC" firstStartedPulling="2025-11-28 14:45:15.408250265 +0000 UTC m=+937.996228531" lastFinishedPulling="2025-11-28 14:45:27.40296897 +0000 UTC m=+949.990947236" observedRunningTime="2025-11-28 14:45:27.879168003 +0000 UTC m=+950.467146279" watchObservedRunningTime="2025-11-28 14:45:27.890295751 +0000 UTC m=+950.478274047" Nov 28 14:45:27 crc kubenswrapper[4817]: I1128 14:45:27.916496 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=18.241403921 podStartE2EDuration="29.916476903s" podCreationTimestamp="2025-11-28 14:44:58 +0000 UTC" firstStartedPulling="2025-11-28 14:45:15.727099088 +0000 UTC m=+938.315077354" lastFinishedPulling="2025-11-28 14:45:27.40217207 +0000 UTC m=+949.990150336" observedRunningTime="2025-11-28 14:45:27.908379091 +0000 UTC m=+950.496357357" watchObservedRunningTime="2025-11-28 14:45:27.916476903 +0000 UTC m=+950.504455189" Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.864554 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p248n" event={"ID":"34bea336-efd1-4432-8ccf-afc87bee4381","Type":"ContainerStarted","Data":"5d9b5cbf93e0d110e77261cb0887cc33589520d002b9e38b780b91a967738a68"} Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.866594 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.867822 4817 generic.go:334] "Generic (PLEG): container finished" podID="ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d" containerID="d4c8acb4f4ca076ef548ee41a75a7e89b594e0bd8220ccd9a8bc20e38b254cf2" exitCode=0 Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.868118 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d","Type":"ContainerDied","Data":"d4c8acb4f4ca076ef548ee41a75a7e89b594e0bd8220ccd9a8bc20e38b254cf2"} Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.871755 4817 generic.go:334] "Generic (PLEG): container finished" podID="bd723f67-725a-4942-87f2-c646eddcd172" containerID="9c75bcc23dbb7e2a619a7f66c20bc7da173287c02298c626f12505394fe53fa5" exitCode=0 Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.871950 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bd723f67-725a-4942-87f2-c646eddcd172","Type":"ContainerDied","Data":"9c75bcc23dbb7e2a619a7f66c20bc7da173287c02298c626f12505394fe53fa5"} Nov 28 14:45:28 crc kubenswrapper[4817]: I1128 14:45:28.892321 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-p248n" podStartSLOduration=2.313478252 podStartE2EDuration="43.892298902s" podCreationTimestamp="2025-11-28 14:44:45 +0000 UTC" firstStartedPulling="2025-11-28 14:44:45.964744153 +0000 UTC m=+908.552722419" lastFinishedPulling="2025-11-28 14:45:27.543564803 +0000 UTC m=+950.131543069" observedRunningTime="2025-11-28 14:45:28.883917613 +0000 UTC m=+951.471895889" watchObservedRunningTime="2025-11-28 14:45:28.892298902 +0000 UTC m=+951.480277178" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.030952 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.102482 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.583455 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bps4t"] Nov 28 14:45:29 crc kubenswrapper[4817]: E1128 14:45:29.583848 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e59a9f-fb61-4364-ace0-8940f22583cb" containerName="collect-profiles" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.583871 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e59a9f-fb61-4364-ace0-8940f22583cb" containerName="collect-profiles" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.584072 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e59a9f-fb61-4364-ace0-8940f22583cb" containerName="collect-profiles" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.585357 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.595319 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bps4t"] Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.615206 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.615302 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.665134 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.741658 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mfgr\" (UniqueName: \"kubernetes.io/projected/43d670c6-8737-40ec-9a71-f8f7347a0c97-kube-api-access-7mfgr\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.742112 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-catalog-content\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.742139 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-utilities\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.843645 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mfgr\" (UniqueName: \"kubernetes.io/projected/43d670c6-8737-40ec-9a71-f8f7347a0c97-kube-api-access-7mfgr\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.843754 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-catalog-content\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.843782 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-utilities\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.844377 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-catalog-content\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.844612 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-utilities\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.865222 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mfgr\" (UniqueName: \"kubernetes.io/projected/43d670c6-8737-40ec-9a71-f8f7347a0c97-kube-api-access-7mfgr\") pod \"redhat-marketplace-bps4t\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.879371 4817 generic.go:334] "Generic (PLEG): container finished" podID="bd260782-cb7b-446c-8222-e28050effe0b" containerID="5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9" exitCode=0 Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.879497 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" event={"ID":"bd260782-cb7b-446c-8222-e28050effe0b","Type":"ContainerDied","Data":"5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9"} Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.880049 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.921447 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 28 14:45:29 crc kubenswrapper[4817]: I1128 14:45:29.924872 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.005986 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.049136 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.225072 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p248n"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.273600 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7kwgh"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.274945 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.277027 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.309329 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7kwgh"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.358486 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.358556 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkn8b\" (UniqueName: \"kubernetes.io/projected/5de0b527-c06b-4250-854b-55c361288d1d-kube-api-access-pkn8b\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.358602 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-config\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.358657 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.398203 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9frjn"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.399299 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.402114 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.440029 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9frjn"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.460503 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.460577 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.460609 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkn8b\" (UniqueName: \"kubernetes.io/projected/5de0b527-c06b-4250-854b-55c361288d1d-kube-api-access-pkn8b\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.460647 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-config\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.464338 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-config\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.464511 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.469971 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.523603 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkn8b\" (UniqueName: \"kubernetes.io/projected/5de0b527-c06b-4250-854b-55c361288d1d-kube-api-access-pkn8b\") pod \"dnsmasq-dns-7fd796d7df-7kwgh\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.563561 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfcl5\" (UniqueName: \"kubernetes.io/projected/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-kube-api-access-mfcl5\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.563978 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.564018 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-config\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.564044 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-ovs-rundir\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.564105 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-ovn-rundir\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.564146 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-combined-ca-bundle\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.575559 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.594855 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.598629 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.598847 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.609611 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-69vsx" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.609833 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.611046 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.619784 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2gk4w"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.625882 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.634363 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bps4t"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665059 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfcl5\" (UniqueName: \"kubernetes.io/projected/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-kube-api-access-mfcl5\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665107 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665146 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2x28\" (UniqueName: \"kubernetes.io/projected/693b7740-a0c9-4430-941a-3f849c254803-kube-api-access-z2x28\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665173 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665206 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-config\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665235 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665259 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-ovs-rundir\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665314 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-ovn-rundir\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665342 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/693b7740-a0c9-4430-941a-3f849c254803-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665367 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665390 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693b7740-a0c9-4430-941a-3f849c254803-config\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665429 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-combined-ca-bundle\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.665450 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693b7740-a0c9-4430-941a-3f849c254803-scripts\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.666465 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-ovs-rundir\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.667030 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-config\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.667090 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-ovn-rundir\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.673127 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-combined-ca-bundle\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.697533 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.726784 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qchx9"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.728232 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.734634 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.735367 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfcl5\" (UniqueName: \"kubernetes.io/projected/9ac085c7-55fa-4dd9-9ab6-bfd720716c60-kube-api-access-mfcl5\") pod \"ovn-controller-metrics-9frjn\" (UID: \"9ac085c7-55fa-4dd9-9ab6-bfd720716c60\") " pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.752741 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qchx9"] Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.765985 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9frjn" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766575 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/693b7740-a0c9-4430-941a-3f849c254803-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766621 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766641 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693b7740-a0c9-4430-941a-3f849c254803-config\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766670 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693b7740-a0c9-4430-941a-3f849c254803-scripts\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766735 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2x28\" (UniqueName: \"kubernetes.io/projected/693b7740-a0c9-4430-941a-3f849c254803-kube-api-access-z2x28\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766756 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.766784 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.767876 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/693b7740-a0c9-4430-941a-3f849c254803-scripts\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.768392 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/693b7740-a0c9-4430-941a-3f849c254803-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.768591 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693b7740-a0c9-4430-941a-3f849c254803-config\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.778219 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.780917 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.786449 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/693b7740-a0c9-4430-941a-3f849c254803-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.800495 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2x28\" (UniqueName: \"kubernetes.io/projected/693b7740-a0c9-4430-941a-3f849c254803-kube-api-access-z2x28\") pod \"ovn-northd-0\" (UID: \"693b7740-a0c9-4430-941a-3f849c254803\") " pod="openstack/ovn-northd-0" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.867799 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.867849 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85mrx\" (UniqueName: \"kubernetes.io/projected/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-kube-api-access-85mrx\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.867869 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-config\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.867929 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.867989 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.903851 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bps4t" event={"ID":"43d670c6-8737-40ec-9a71-f8f7347a0c97","Type":"ContainerStarted","Data":"f492db1a2c5aface0ddab77845cb54b022d6a0b4320d71c55a659fe0fc2487c2"} Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.904337 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-p248n" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="dnsmasq-dns" containerID="cri-o://5d9b5cbf93e0d110e77261cb0887cc33589520d002b9e38b780b91a967738a68" gracePeriod=10 Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.968873 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.968949 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85mrx\" (UniqueName: \"kubernetes.io/projected/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-kube-api-access-85mrx\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.968967 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-config\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.969018 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.969101 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.970561 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-config\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.971250 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.971847 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:30 crc kubenswrapper[4817]: I1128 14:45:30.975213 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.005456 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85mrx\" (UniqueName: \"kubernetes.io/projected/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-kube-api-access-85mrx\") pod \"dnsmasq-dns-86db49b7ff-qchx9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.062552 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.089816 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.288883 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7kwgh"] Nov 28 14:45:31 crc kubenswrapper[4817]: W1128 14:45:31.293427 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5de0b527_c06b_4250_854b_55c361288d1d.slice/crio-28e83a4c7c85d918636ba72a4a9973483f33225bfecd6b24920d7e5cd581790b WatchSource:0}: Error finding container 28e83a4c7c85d918636ba72a4a9973483f33225bfecd6b24920d7e5cd581790b: Status 404 returned error can't find the container with id 28e83a4c7c85d918636ba72a4a9973483f33225bfecd6b24920d7e5cd581790b Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.383850 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9frjn"] Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.615597 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 14:45:31 crc kubenswrapper[4817]: W1128 14:45:31.616256 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod693b7740_a0c9_4430_941a_3f849c254803.slice/crio-a96f2843c0e08ba37396530f29b2e536cc418ee1b2d1d40c6494c6370a730f53 WatchSource:0}: Error finding container a96f2843c0e08ba37396530f29b2e536cc418ee1b2d1d40c6494c6370a730f53: Status 404 returned error can't find the container with id a96f2843c0e08ba37396530f29b2e536cc418ee1b2d1d40c6494c6370a730f53 Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.670477 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qchx9"] Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.911439 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9frjn" event={"ID":"9ac085c7-55fa-4dd9-9ab6-bfd720716c60","Type":"ContainerStarted","Data":"ced9c68d97f79a1ebfae054f3cb0af1b2daaa9212e7be3bf4ac758c16cb82ddb"} Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.912777 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" event={"ID":"5de0b527-c06b-4250-854b-55c361288d1d","Type":"ContainerStarted","Data":"28e83a4c7c85d918636ba72a4a9973483f33225bfecd6b24920d7e5cd581790b"} Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.914047 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" event={"ID":"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9","Type":"ContainerStarted","Data":"67b911ea1d230ae5d572c17a6c603048f43d2af69ca2e20d0e7c9fd98fee5581"} Nov 28 14:45:31 crc kubenswrapper[4817]: I1128 14:45:31.915763 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"693b7740-a0c9-4430-941a-3f849c254803","Type":"ContainerStarted","Data":"a96f2843c0e08ba37396530f29b2e536cc418ee1b2d1d40c6494c6370a730f53"} Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.033468 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7kwgh"] Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.086049 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-pr6vs"] Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.087673 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.099057 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.105397 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pr6vs"] Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.196832 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.196917 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8g5c\" (UniqueName: \"kubernetes.io/projected/6586dc95-7ca6-476a-a7d7-67e550cb19e9-kube-api-access-r8g5c\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.196981 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.197028 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-dns-svc\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.197055 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-config\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.298919 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-dns-svc\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.298979 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-config\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.299078 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.299126 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8g5c\" (UniqueName: \"kubernetes.io/projected/6586dc95-7ca6-476a-a7d7-67e550cb19e9-kube-api-access-r8g5c\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.299152 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.299808 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-dns-svc\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.300099 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.300182 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-config\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.300247 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.317710 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8g5c\" (UniqueName: \"kubernetes.io/projected/6586dc95-7ca6-476a-a7d7-67e550cb19e9-kube-api-access-r8g5c\") pod \"dnsmasq-dns-698758b865-pr6vs\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.403851 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.896936 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pr6vs"] Nov 28 14:45:32 crc kubenswrapper[4817]: W1128 14:45:32.905608 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6586dc95_7ca6_476a_a7d7_67e550cb19e9.slice/crio-a2ad9a39b4f592c38c69da82daa18719dbe38fb58d98d0b02903c4f2dc08c957 WatchSource:0}: Error finding container a2ad9a39b4f592c38c69da82daa18719dbe38fb58d98d0b02903c4f2dc08c957: Status 404 returned error can't find the container with id a2ad9a39b4f592c38c69da82daa18719dbe38fb58d98d0b02903c4f2dc08c957 Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.924298 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d","Type":"ContainerStarted","Data":"77ca8f79c993515786e83ae565f7c76665b7c4c6c427f1f0fb66643f0b263765"} Nov 28 14:45:32 crc kubenswrapper[4817]: I1128 14:45:32.925301 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pr6vs" event={"ID":"6586dc95-7ca6-476a-a7d7-67e550cb19e9","Type":"ContainerStarted","Data":"a2ad9a39b4f592c38c69da82daa18719dbe38fb58d98d0b02903c4f2dc08c957"} Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.157429 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.157800 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.216085 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.229391 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.235046 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.237033 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.237251 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.237286 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-njl4h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.237835 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.255375 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.315629 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b674c\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-kube-api-access-b674c\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.316031 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.316166 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/42276c9f-fac8-4794-a69c-02178767e5b1-cache\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.316293 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.316428 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/42276c9f-fac8-4794-a69c-02178767e5b1-lock\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418026 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/42276c9f-fac8-4794-a69c-02178767e5b1-lock\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418159 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b674c\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-kube-api-access-b674c\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418230 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418268 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/42276c9f-fac8-4794-a69c-02178767e5b1-cache\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418314 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418555 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/42276c9f-fac8-4794-a69c-02178767e5b1-lock\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.418558 4817 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.418620 4817 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.418673 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift podName:42276c9f-fac8-4794-a69c-02178767e5b1 nodeName:}" failed. No retries permitted until 2025-11-28 14:45:33.918646413 +0000 UTC m=+956.506624679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift") pod "swift-storage-0" (UID: "42276c9f-fac8-4794-a69c-02178767e5b1") : configmap "swift-ring-files" not found Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418710 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/42276c9f-fac8-4794-a69c-02178767e5b1-cache\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.418812 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.771291 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6vl8h"] Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.772926 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.776553 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.776787 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.776818 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.793162 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6vl8h"] Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.842186 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-ring-data-devices\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.842308 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d6d1bafc-477f-4e90-996c-b3fe1093ba16-etc-swift\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.843513 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-combined-ca-bundle\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.843608 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-dispersionconf\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.845322 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-scripts\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.845462 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzvhr\" (UniqueName: \"kubernetes.io/projected/d6d1bafc-477f-4e90-996c-b3fe1093ba16-kube-api-access-gzvhr\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.845583 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-swiftconf\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.867890 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6vl8h"] Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.869351 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-gzvhr ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-6vl8h" podUID="d6d1bafc-477f-4e90-996c-b3fe1093ba16" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.884447 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-fzhb6"] Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.885644 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.892279 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fzhb6"] Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.934264 4817 generic.go:334] "Generic (PLEG): container finished" podID="34bea336-efd1-4432-8ccf-afc87bee4381" containerID="5d9b5cbf93e0d110e77261cb0887cc33589520d002b9e38b780b91a967738a68" exitCode=0 Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.935104 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p248n" event={"ID":"34bea336-efd1-4432-8ccf-afc87bee4381","Type":"ContainerDied","Data":"5d9b5cbf93e0d110e77261cb0887cc33589520d002b9e38b780b91a967738a68"} Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.935159 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.943837 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.948905 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-scripts\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.948965 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-combined-ca-bundle\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.948994 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-etc-swift\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949028 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-dispersionconf\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949069 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-ring-data-devices\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949103 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-scripts\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949171 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzvhr\" (UniqueName: \"kubernetes.io/projected/d6d1bafc-477f-4e90-996c-b3fe1093ba16-kube-api-access-gzvhr\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949199 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-combined-ca-bundle\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949239 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khlcc\" (UniqueName: \"kubernetes.io/projected/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-kube-api-access-khlcc\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949280 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-dispersionconf\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949304 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-swiftconf\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949337 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949377 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-ring-data-devices\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949416 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d6d1bafc-477f-4e90-996c-b3fe1093ba16-etc-swift\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.949461 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-swiftconf\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.949950 4817 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.949972 4817 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 14:45:33 crc kubenswrapper[4817]: E1128 14:45:33.950029 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift podName:42276c9f-fac8-4794-a69c-02178767e5b1 nodeName:}" failed. No retries permitted until 2025-11-28 14:45:34.950014741 +0000 UTC m=+957.537993007 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift") pod "swift-storage-0" (UID: "42276c9f-fac8-4794-a69c-02178767e5b1") : configmap "swift-ring-files" not found Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.950328 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-scripts\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.950742 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d6d1bafc-477f-4e90-996c-b3fe1093ba16-etc-swift\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.950886 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-ring-data-devices\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.953851 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-swiftconf\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.953988 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-dispersionconf\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.966494 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzvhr\" (UniqueName: \"kubernetes.io/projected/d6d1bafc-477f-4e90-996c-b3fe1093ba16-kube-api-access-gzvhr\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.977550 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-combined-ca-bundle\") pod \"swift-ring-rebalance-6vl8h\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:33 crc kubenswrapper[4817]: I1128 14:45:33.981859 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.021011 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-blkhg"] Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.051061 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-scripts\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.051104 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d6d1bafc-477f-4e90-996c-b3fe1093ba16-etc-swift\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.051601 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6d1bafc-477f-4e90-996c-b3fe1093ba16-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.051673 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzvhr\" (UniqueName: \"kubernetes.io/projected/d6d1bafc-477f-4e90-996c-b3fe1093ba16-kube-api-access-gzvhr\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.051736 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-scripts" (OuterVolumeSpecName: "scripts") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052232 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-swiftconf\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052289 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-dispersionconf\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052339 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-ring-data-devices\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052355 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-combined-ca-bundle\") pod \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\" (UID: \"d6d1bafc-477f-4e90-996c-b3fe1093ba16\") " Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052666 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-swiftconf\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052695 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-scripts\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052737 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-etc-swift\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.052771 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-ring-data-devices\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.053004 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-combined-ca-bundle\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.053030 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khlcc\" (UniqueName: \"kubernetes.io/projected/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-kube-api-access-khlcc\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.053061 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-dispersionconf\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.053103 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.053114 4817 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d6d1bafc-477f-4e90-996c-b3fe1093ba16-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.054607 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-scripts\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.054656 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-etc-swift\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.055005 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.055024 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-ring-data-devices\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.060499 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d1bafc-477f-4e90-996c-b3fe1093ba16-kube-api-access-gzvhr" (OuterVolumeSpecName: "kube-api-access-gzvhr") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "kube-api-access-gzvhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.071075 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.073735 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-combined-ca-bundle\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.074329 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-dispersionconf\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.076937 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.077383 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-swiftconf\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.077899 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d6d1bafc-477f-4e90-996c-b3fe1093ba16" (UID: "d6d1bafc-477f-4e90-996c-b3fe1093ba16"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.089265 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khlcc\" (UniqueName: \"kubernetes.io/projected/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-kube-api-access-khlcc\") pod \"swift-ring-rebalance-fzhb6\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.121891 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b674c\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-kube-api-access-b674c\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.123271 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.155322 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzvhr\" (UniqueName: \"kubernetes.io/projected/d6d1bafc-477f-4e90-996c-b3fe1093ba16-kube-api-access-gzvhr\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.155374 4817 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.155386 4817 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.155398 4817 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d6d1bafc-477f-4e90-996c-b3fe1093ba16-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.155425 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d1bafc-477f-4e90-996c-b3fe1093ba16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.205636 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.695701 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fzhb6"] Nov 28 14:45:34 crc kubenswrapper[4817]: W1128 14:45:34.700522 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf61aa2a2_1ba8_4319_ad02_2c05d6a9f662.slice/crio-1cd7822c724de04b6cc4dcf022fa2dc75926152d577d7478166b8c59e0828d3d WatchSource:0}: Error finding container 1cd7822c724de04b6cc4dcf022fa2dc75926152d577d7478166b8c59e0828d3d: Status 404 returned error can't find the container with id 1cd7822c724de04b6cc4dcf022fa2dc75926152d577d7478166b8c59e0828d3d Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.941714 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fzhb6" event={"ID":"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662","Type":"ContainerStarted","Data":"1cd7822c724de04b6cc4dcf022fa2dc75926152d577d7478166b8c59e0828d3d"} Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.941794 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6vl8h" Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.970506 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:34 crc kubenswrapper[4817]: E1128 14:45:34.970848 4817 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 14:45:34 crc kubenswrapper[4817]: E1128 14:45:34.970881 4817 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 14:45:34 crc kubenswrapper[4817]: E1128 14:45:34.970945 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift podName:42276c9f-fac8-4794-a69c-02178767e5b1 nodeName:}" failed. No retries permitted until 2025-11-28 14:45:36.970922825 +0000 UTC m=+959.558901111 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift") pod "swift-storage-0" (UID: "42276c9f-fac8-4794-a69c-02178767e5b1") : configmap "swift-ring-files" not found Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.987803 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6vl8h"] Nov 28 14:45:34 crc kubenswrapper[4817]: I1128 14:45:34.997103 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-6vl8h"] Nov 28 14:45:35 crc kubenswrapper[4817]: I1128 14:45:35.354827 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-p248n" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.91:5353: connect: connection refused" Nov 28 14:45:35 crc kubenswrapper[4817]: I1128 14:45:35.748061 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6d1bafc-477f-4e90-996c-b3fe1093ba16" path="/var/lib/kubelet/pods/d6d1bafc-477f-4e90-996c-b3fe1093ba16/volumes" Nov 28 14:45:35 crc kubenswrapper[4817]: I1128 14:45:35.953248 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-blkhg" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="registry-server" containerID="cri-o://e0f32889994eb0364eca54cf0dd12567a88917189596c669f745f4fc455c1611" gracePeriod=2 Nov 28 14:45:36 crc kubenswrapper[4817]: I1128 14:45:36.964254 4817 generic.go:334] "Generic (PLEG): container finished" podID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerID="e0f32889994eb0364eca54cf0dd12567a88917189596c669f745f4fc455c1611" exitCode=0 Nov 28 14:45:36 crc kubenswrapper[4817]: I1128 14:45:36.964331 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blkhg" event={"ID":"c3d3ce03-7abf-4cd7-8394-192f8c56ba51","Type":"ContainerDied","Data":"e0f32889994eb0364eca54cf0dd12567a88917189596c669f745f4fc455c1611"} Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.002565 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:37 crc kubenswrapper[4817]: E1128 14:45:37.002768 4817 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 14:45:37 crc kubenswrapper[4817]: E1128 14:45:37.002792 4817 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 14:45:37 crc kubenswrapper[4817]: E1128 14:45:37.002855 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift podName:42276c9f-fac8-4794-a69c-02178767e5b1 nodeName:}" failed. No retries permitted until 2025-11-28 14:45:41.002834144 +0000 UTC m=+963.590812410 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift") pod "swift-storage-0" (UID: "42276c9f-fac8-4794-a69c-02178767e5b1") : configmap "swift-ring-files" not found Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.560175 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.617234 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w28r6\" (UniqueName: \"kubernetes.io/projected/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-kube-api-access-w28r6\") pod \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.617431 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-catalog-content\") pod \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.617505 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-utilities\") pod \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\" (UID: \"c3d3ce03-7abf-4cd7-8394-192f8c56ba51\") " Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.618948 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-utilities" (OuterVolumeSpecName: "utilities") pod "c3d3ce03-7abf-4cd7-8394-192f8c56ba51" (UID: "c3d3ce03-7abf-4cd7-8394-192f8c56ba51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.624609 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-kube-api-access-w28r6" (OuterVolumeSpecName: "kube-api-access-w28r6") pod "c3d3ce03-7abf-4cd7-8394-192f8c56ba51" (UID: "c3d3ce03-7abf-4cd7-8394-192f8c56ba51"). InnerVolumeSpecName "kube-api-access-w28r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.720434 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.720820 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w28r6\" (UniqueName: \"kubernetes.io/projected/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-kube-api-access-w28r6\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.976214 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blkhg" event={"ID":"c3d3ce03-7abf-4cd7-8394-192f8c56ba51","Type":"ContainerDied","Data":"2849bcba6b796f8c8c25ecbacbc185fc1005c347d1875ef33d76c83b98e53cbb"} Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.976271 4817 scope.go:117] "RemoveContainer" containerID="e0f32889994eb0364eca54cf0dd12567a88917189596c669f745f4fc455c1611" Nov 28 14:45:37 crc kubenswrapper[4817]: I1128 14:45:37.976314 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blkhg" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.002923 4817 scope.go:117] "RemoveContainer" containerID="883a32419ad50c9bc2eb8dbe28886a53d9e7db848731eb8d66cfca1f1019eefb" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.019181 4817 scope.go:117] "RemoveContainer" containerID="f704860adb9476fc82697e701a70dbc79a26b7b31cd0d05f9241a86e359d8a03" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.276362 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3d3ce03-7abf-4cd7-8394-192f8c56ba51" (UID: "c3d3ce03-7abf-4cd7-8394-192f8c56ba51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.333003 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d3ce03-7abf-4cd7-8394-192f8c56ba51-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.893998 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.910794 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-blkhg"] Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.934893 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-blkhg"] Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.950283 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gmgf\" (UniqueName: \"kubernetes.io/projected/34bea336-efd1-4432-8ccf-afc87bee4381-kube-api-access-7gmgf\") pod \"34bea336-efd1-4432-8ccf-afc87bee4381\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.950491 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-dns-svc\") pod \"34bea336-efd1-4432-8ccf-afc87bee4381\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.950563 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-config\") pod \"34bea336-efd1-4432-8ccf-afc87bee4381\" (UID: \"34bea336-efd1-4432-8ccf-afc87bee4381\") " Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.957166 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34bea336-efd1-4432-8ccf-afc87bee4381-kube-api-access-7gmgf" (OuterVolumeSpecName: "kube-api-access-7gmgf") pod "34bea336-efd1-4432-8ccf-afc87bee4381" (UID: "34bea336-efd1-4432-8ccf-afc87bee4381"). InnerVolumeSpecName "kube-api-access-7gmgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:38 crc kubenswrapper[4817]: I1128 14:45:38.999048 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9frjn" event={"ID":"9ac085c7-55fa-4dd9-9ab6-bfd720716c60","Type":"ContainerStarted","Data":"8f21ab7f5b1fc2aa83d169c44a402010562c6595b7aa8571ae61d89679e43a55"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.000879 4817 generic.go:334] "Generic (PLEG): container finished" podID="5de0b527-c06b-4250-854b-55c361288d1d" containerID="561cf6fd38a9beedded8195dd4cdd57ae3ab6f623516607da0885436d3b9e709" exitCode=0 Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.000936 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" event={"ID":"5de0b527-c06b-4250-854b-55c361288d1d","Type":"ContainerDied","Data":"561cf6fd38a9beedded8195dd4cdd57ae3ab6f623516607da0885436d3b9e709"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.002845 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-p248n" event={"ID":"34bea336-efd1-4432-8ccf-afc87bee4381","Type":"ContainerDied","Data":"c007ae040ebc42bf71e9916f77c3edd74935748098bfe580641b429284558e4e"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.002891 4817 scope.go:117] "RemoveContainer" containerID="5d9b5cbf93e0d110e77261cb0887cc33589520d002b9e38b780b91a967738a68" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.002887 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-p248n" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.005574 4817 generic.go:334] "Generic (PLEG): container finished" podID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerID="5335d7c7f38b3ce4ca3b2fa4985cce99d73208bfc7870db008f21dbf080b7efa" exitCode=0 Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.005602 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pr6vs" event={"ID":"6586dc95-7ca6-476a-a7d7-67e550cb19e9","Type":"ContainerDied","Data":"5335d7c7f38b3ce4ca3b2fa4985cce99d73208bfc7870db008f21dbf080b7efa"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.008655 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-config" (OuterVolumeSpecName: "config") pod "34bea336-efd1-4432-8ccf-afc87bee4381" (UID: "34bea336-efd1-4432-8ccf-afc87bee4381"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.009066 4817 generic.go:334] "Generic (PLEG): container finished" podID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerID="facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d" exitCode=0 Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.009141 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bps4t" event={"ID":"43d670c6-8737-40ec-9a71-f8f7347a0c97","Type":"ContainerDied","Data":"facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.011747 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" event={"ID":"bd260782-cb7b-446c-8222-e28050effe0b","Type":"ContainerStarted","Data":"c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.012663 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" podUID="bd260782-cb7b-446c-8222-e28050effe0b" containerName="dnsmasq-dns" containerID="cri-o://c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665" gracePeriod=10 Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.012697 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.021602 4817 generic.go:334] "Generic (PLEG): container finished" podID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerID="afafef7f7b2e467ef846e03f9647812e13d1f44a510620cc4cb229ad63e8762d" exitCode=0 Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.021706 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" event={"ID":"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9","Type":"ContainerDied","Data":"afafef7f7b2e467ef846e03f9647812e13d1f44a510620cc4cb229ad63e8762d"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.022442 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34bea336-efd1-4432-8ccf-afc87bee4381" (UID: "34bea336-efd1-4432-8ccf-afc87bee4381"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.023984 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9frjn" podStartSLOduration=9.023943604 podStartE2EDuration="9.023943604s" podCreationTimestamp="2025-11-28 14:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:45:39.016044338 +0000 UTC m=+961.604022604" watchObservedRunningTime="2025-11-28 14:45:39.023943604 +0000 UTC m=+961.611921870" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.031689 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bd723f67-725a-4942-87f2-c646eddcd172","Type":"ContainerStarted","Data":"296b6195a671b50d72eb9a0c2035387d2d99e64e1179806074f49c8da28c47c3"} Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.045465 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" podStartSLOduration=-9223371982.809336 podStartE2EDuration="54.04544105s" podCreationTimestamp="2025-11-28 14:44:45 +0000 UTC" firstStartedPulling="2025-11-28 14:44:46.236170634 +0000 UTC m=+908.824148910" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:45:39.038460016 +0000 UTC m=+961.626438292" watchObservedRunningTime="2025-11-28 14:45:39.04544105 +0000 UTC m=+961.633419306" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.060146 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.060196 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34bea336-efd1-4432-8ccf-afc87bee4381-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.060357 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gmgf\" (UniqueName: \"kubernetes.io/projected/34bea336-efd1-4432-8ccf-afc87bee4381-kube-api-access-7gmgf\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.168844 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=43.757758088 podStartE2EDuration="51.168828674s" podCreationTimestamp="2025-11-28 14:44:48 +0000 UTC" firstStartedPulling="2025-11-28 14:45:15.424335196 +0000 UTC m=+938.012313462" lastFinishedPulling="2025-11-28 14:45:22.835405782 +0000 UTC m=+945.423384048" observedRunningTime="2025-11-28 14:45:39.167943272 +0000 UTC m=+961.755921538" watchObservedRunningTime="2025-11-28 14:45:39.168828674 +0000 UTC m=+961.756806940" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.217500 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=45.173399539 podStartE2EDuration="53.217479485s" podCreationTimestamp="2025-11-28 14:44:46 +0000 UTC" firstStartedPulling="2025-11-28 14:45:14.746459078 +0000 UTC m=+937.334437344" lastFinishedPulling="2025-11-28 14:45:22.790539024 +0000 UTC m=+945.378517290" observedRunningTime="2025-11-28 14:45:39.212572893 +0000 UTC m=+961.800551159" watchObservedRunningTime="2025-11-28 14:45:39.217479485 +0000 UTC m=+961.805457751" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.340211 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p248n"] Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.347762 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-p248n"] Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.452049 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.468819 4817 scope.go:117] "RemoveContainer" containerID="2e28bf49af60e63d337794899a865b6df304e2e1d53f2d5fea9aa1a058ab7e7d" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.568802 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-dns-svc\") pod \"5de0b527-c06b-4250-854b-55c361288d1d\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.568848 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-ovsdbserver-nb\") pod \"5de0b527-c06b-4250-854b-55c361288d1d\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.568985 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-config\") pod \"5de0b527-c06b-4250-854b-55c361288d1d\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.569041 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkn8b\" (UniqueName: \"kubernetes.io/projected/5de0b527-c06b-4250-854b-55c361288d1d-kube-api-access-pkn8b\") pod \"5de0b527-c06b-4250-854b-55c361288d1d\" (UID: \"5de0b527-c06b-4250-854b-55c361288d1d\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.573647 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de0b527-c06b-4250-854b-55c361288d1d-kube-api-access-pkn8b" (OuterVolumeSpecName: "kube-api-access-pkn8b") pod "5de0b527-c06b-4250-854b-55c361288d1d" (UID: "5de0b527-c06b-4250-854b-55c361288d1d"). InnerVolumeSpecName "kube-api-access-pkn8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.591569 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-config" (OuterVolumeSpecName: "config") pod "5de0b527-c06b-4250-854b-55c361288d1d" (UID: "5de0b527-c06b-4250-854b-55c361288d1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.592319 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5de0b527-c06b-4250-854b-55c361288d1d" (UID: "5de0b527-c06b-4250-854b-55c361288d1d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.593542 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5de0b527-c06b-4250-854b-55c361288d1d" (UID: "5de0b527-c06b-4250-854b-55c361288d1d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.641828 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.641886 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.671485 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.671574 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkn8b\" (UniqueName: \"kubernetes.io/projected/5de0b527-c06b-4250-854b-55c361288d1d-kube-api-access-pkn8b\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.671594 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.671608 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5de0b527-c06b-4250-854b-55c361288d1d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.700615 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.747925 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" path="/var/lib/kubelet/pods/34bea336-efd1-4432-8ccf-afc87bee4381/volumes" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.748782 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" path="/var/lib/kubelet/pods/c3d3ce03-7abf-4cd7-8394-192f8c56ba51/volumes" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.773228 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-dns-svc\") pod \"bd260782-cb7b-446c-8222-e28050effe0b\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.773336 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-config\") pod \"bd260782-cb7b-446c-8222-e28050effe0b\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.773480 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcv4v\" (UniqueName: \"kubernetes.io/projected/bd260782-cb7b-446c-8222-e28050effe0b-kube-api-access-zcv4v\") pod \"bd260782-cb7b-446c-8222-e28050effe0b\" (UID: \"bd260782-cb7b-446c-8222-e28050effe0b\") " Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.784770 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd260782-cb7b-446c-8222-e28050effe0b-kube-api-access-zcv4v" (OuterVolumeSpecName: "kube-api-access-zcv4v") pod "bd260782-cb7b-446c-8222-e28050effe0b" (UID: "bd260782-cb7b-446c-8222-e28050effe0b"). InnerVolumeSpecName "kube-api-access-zcv4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.831933 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-config" (OuterVolumeSpecName: "config") pod "bd260782-cb7b-446c-8222-e28050effe0b" (UID: "bd260782-cb7b-446c-8222-e28050effe0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.835202 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd260782-cb7b-446c-8222-e28050effe0b" (UID: "bd260782-cb7b-446c-8222-e28050effe0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.878032 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.878063 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd260782-cb7b-446c-8222-e28050effe0b-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:39 crc kubenswrapper[4817]: I1128 14:45:39.878073 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcv4v\" (UniqueName: \"kubernetes.io/projected/bd260782-cb7b-446c-8222-e28050effe0b-kube-api-access-zcv4v\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:40 crc kubenswrapper[4817]: E1128 14:45:40.023152 4817 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.236:59364->38.102.83.236:36927: write tcp 38.102.83.236:59364->38.102.83.236:36927: write: broken pipe Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.040755 4817 generic.go:334] "Generic (PLEG): container finished" podID="bd260782-cb7b-446c-8222-e28050effe0b" containerID="c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665" exitCode=0 Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.040823 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.040833 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" event={"ID":"bd260782-cb7b-446c-8222-e28050effe0b","Type":"ContainerDied","Data":"c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665"} Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.040867 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2gk4w" event={"ID":"bd260782-cb7b-446c-8222-e28050effe0b","Type":"ContainerDied","Data":"4f17be5e5bd9cf44bfa675c7cdeb089694b2b950cd0387e0a944b52df64132a8"} Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.040892 4817 scope.go:117] "RemoveContainer" containerID="c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665" Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.056014 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"693b7740-a0c9-4430-941a-3f849c254803","Type":"ContainerStarted","Data":"6a28d64909e7993ac98c2a7818620390c1fd5496d89a0b55c5b82e3fb0d546e2"} Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.061543 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pr6vs" event={"ID":"6586dc95-7ca6-476a-a7d7-67e550cb19e9","Type":"ContainerStarted","Data":"3f32a9dec98e4bfd02513cbbcf1005b69f8f8798fb6b0d6c710039c661bcec02"} Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.061790 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.063755 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.063781 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7kwgh" event={"ID":"5de0b527-c06b-4250-854b-55c361288d1d","Type":"ContainerDied","Data":"28e83a4c7c85d918636ba72a4a9973483f33225bfecd6b24920d7e5cd581790b"} Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.077774 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" event={"ID":"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9","Type":"ContainerStarted","Data":"16e17218abb17dc509615f4bbde4f575898ac43de3b6626ce6201c990d1d7bcb"} Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.079276 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2gk4w"] Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.085634 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2gk4w"] Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.100829 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-pr6vs" podStartSLOduration=8.10078418 podStartE2EDuration="8.10078418s" podCreationTimestamp="2025-11-28 14:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:45:40.091473448 +0000 UTC m=+962.679451714" watchObservedRunningTime="2025-11-28 14:45:40.10078418 +0000 UTC m=+962.688762446" Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.133236 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7kwgh"] Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.140492 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7kwgh"] Nov 28 14:45:40 crc kubenswrapper[4817]: I1128 14:45:40.151990 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" podStartSLOduration=10.151973645 podStartE2EDuration="10.151973645s" podCreationTimestamp="2025-11-28 14:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:45:40.144570241 +0000 UTC m=+962.732548517" watchObservedRunningTime="2025-11-28 14:45:40.151973645 +0000 UTC m=+962.739951911" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.090332 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.099073 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:41 crc kubenswrapper[4817]: E1128 14:45:41.099315 4817 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 14:45:41 crc kubenswrapper[4817]: E1128 14:45:41.099333 4817 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 14:45:41 crc kubenswrapper[4817]: E1128 14:45:41.099393 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift podName:42276c9f-fac8-4794-a69c-02178767e5b1 nodeName:}" failed. No retries permitted until 2025-11-28 14:45:49.099377247 +0000 UTC m=+971.687355513 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift") pod "swift-storage-0" (UID: "42276c9f-fac8-4794-a69c-02178767e5b1") : configmap "swift-ring-files" not found Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.112084 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"693b7740-a0c9-4430-941a-3f849c254803","Type":"ContainerStarted","Data":"694b79f04ecc858de70c2515e85a199592fd617570c4bcc551b8649779374c31"} Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.112609 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.133053 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.208286574 podStartE2EDuration="11.133032586s" podCreationTimestamp="2025-11-28 14:45:30 +0000 UTC" firstStartedPulling="2025-11-28 14:45:31.61895051 +0000 UTC m=+954.206928766" lastFinishedPulling="2025-11-28 14:45:39.543696512 +0000 UTC m=+962.131674778" observedRunningTime="2025-11-28 14:45:41.127228931 +0000 UTC m=+963.715207197" watchObservedRunningTime="2025-11-28 14:45:41.133032586 +0000 UTC m=+963.721010842" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.747540 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de0b527-c06b-4250-854b-55c361288d1d" path="/var/lib/kubelet/pods/5de0b527-c06b-4250-854b-55c361288d1d/volumes" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.750414 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd260782-cb7b-446c-8222-e28050effe0b" path="/var/lib/kubelet/pods/bd260782-cb7b-446c-8222-e28050effe0b/volumes" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.894031 4817 scope.go:117] "RemoveContainer" containerID="5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.914582 4817 scope.go:117] "RemoveContainer" containerID="c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665" Nov 28 14:45:41 crc kubenswrapper[4817]: E1128 14:45:41.915004 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665\": container with ID starting with c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665 not found: ID does not exist" containerID="c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.915044 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665"} err="failed to get container status \"c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665\": rpc error: code = NotFound desc = could not find container \"c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665\": container with ID starting with c8a12d9c8634d44bf5a6c0a303125329cbb4c4f16c9f22350ecb22cc7de43665 not found: ID does not exist" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.915071 4817 scope.go:117] "RemoveContainer" containerID="5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9" Nov 28 14:45:41 crc kubenswrapper[4817]: E1128 14:45:41.915464 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9\": container with ID starting with 5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9 not found: ID does not exist" containerID="5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.915499 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9"} err="failed to get container status \"5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9\": rpc error: code = NotFound desc = could not find container \"5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9\": container with ID starting with 5246706e96d2b15f367e4196f4c477220e31a0b5d04c1e0cbf638144f37542e9 not found: ID does not exist" Nov 28 14:45:41 crc kubenswrapper[4817]: I1128 14:45:41.915518 4817 scope.go:117] "RemoveContainer" containerID="561cf6fd38a9beedded8195dd4cdd57ae3ab6f623516607da0885436d3b9e709" Nov 28 14:45:43 crc kubenswrapper[4817]: I1128 14:45:43.139990 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fzhb6" event={"ID":"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662","Type":"ContainerStarted","Data":"801afa80588778b2ac91cca24026c967004c701c8409f822591f84050b871051"} Nov 28 14:45:43 crc kubenswrapper[4817]: I1128 14:45:43.141731 4817 generic.go:334] "Generic (PLEG): container finished" podID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerID="997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756" exitCode=0 Nov 28 14:45:43 crc kubenswrapper[4817]: I1128 14:45:43.141756 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bps4t" event={"ID":"43d670c6-8737-40ec-9a71-f8f7347a0c97","Type":"ContainerDied","Data":"997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756"} Nov 28 14:45:43 crc kubenswrapper[4817]: I1128 14:45:43.161712 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-fzhb6" podStartSLOduration=2.892116413 podStartE2EDuration="10.161693724s" podCreationTimestamp="2025-11-28 14:45:33 +0000 UTC" firstStartedPulling="2025-11-28 14:45:34.702480627 +0000 UTC m=+957.290458893" lastFinishedPulling="2025-11-28 14:45:41.972057938 +0000 UTC m=+964.560036204" observedRunningTime="2025-11-28 14:45:43.155693324 +0000 UTC m=+965.743671600" watchObservedRunningTime="2025-11-28 14:45:43.161693724 +0000 UTC m=+965.749671990" Nov 28 14:45:43 crc kubenswrapper[4817]: I1128 14:45:43.804397 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 28 14:45:43 crc kubenswrapper[4817]: I1128 14:45:43.886515 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 28 14:45:44 crc kubenswrapper[4817]: I1128 14:45:44.153488 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bps4t" event={"ID":"43d670c6-8737-40ec-9a71-f8f7347a0c97","Type":"ContainerStarted","Data":"9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb"} Nov 28 14:45:44 crc kubenswrapper[4817]: I1128 14:45:44.187167 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bps4t" podStartSLOduration=10.824251571 podStartE2EDuration="15.18715062s" podCreationTimestamp="2025-11-28 14:45:29 +0000 UTC" firstStartedPulling="2025-11-28 14:45:39.43563176 +0000 UTC m=+962.023610026" lastFinishedPulling="2025-11-28 14:45:43.798530809 +0000 UTC m=+966.386509075" observedRunningTime="2025-11-28 14:45:44.180651708 +0000 UTC m=+966.768629974" watchObservedRunningTime="2025-11-28 14:45:44.18715062 +0000 UTC m=+966.775128876" Nov 28 14:45:46 crc kubenswrapper[4817]: I1128 14:45:46.091746 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:47 crc kubenswrapper[4817]: I1128 14:45:47.404883 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:45:47 crc kubenswrapper[4817]: I1128 14:45:47.484758 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qchx9"] Nov 28 14:45:47 crc kubenswrapper[4817]: I1128 14:45:47.485132 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerName="dnsmasq-dns" containerID="cri-o://16e17218abb17dc509615f4bbde4f575898ac43de3b6626ce6201c990d1d7bcb" gracePeriod=10 Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.070886 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.071226 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.185316 4817 generic.go:334] "Generic (PLEG): container finished" podID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerID="290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4" exitCode=0 Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.185409 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63b71752-f06a-4d1a-9d80-85142bbc3610","Type":"ContainerDied","Data":"290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4"} Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.197169 4817 generic.go:334] "Generic (PLEG): container finished" podID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerID="16e17218abb17dc509615f4bbde4f575898ac43de3b6626ce6201c990d1d7bcb" exitCode=0 Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.197215 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" event={"ID":"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9","Type":"ContainerDied","Data":"16e17218abb17dc509615f4bbde4f575898ac43de3b6626ce6201c990d1d7bcb"} Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.417282 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.431540 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.532654 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-nb\") pod \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.532774 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-config\") pod \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.532849 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-dns-svc\") pod \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.532904 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-sb\") pod \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.532935 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85mrx\" (UniqueName: \"kubernetes.io/projected/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-kube-api-access-85mrx\") pod \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\" (UID: \"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9\") " Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.539678 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.540090 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-kube-api-access-85mrx" (OuterVolumeSpecName: "kube-api-access-85mrx") pod "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" (UID: "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9"). InnerVolumeSpecName "kube-api-access-85mrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.594941 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" (UID: "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.604633 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" (UID: "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.614522 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" (UID: "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.634696 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.634738 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.634750 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85mrx\" (UniqueName: \"kubernetes.io/projected/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-kube-api-access-85mrx\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.634758 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.644243 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-config" (OuterVolumeSpecName: "config") pod "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" (UID: "b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:48 crc kubenswrapper[4817]: I1128 14:45:48.736367 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.141687 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.141904 4817 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.141936 4817 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.142005 4817 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift podName:42276c9f-fac8-4794-a69c-02178767e5b1 nodeName:}" failed. No retries permitted until 2025-11-28 14:46:05.141983407 +0000 UTC m=+987.729961673 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift") pod "swift-storage-0" (UID: "42276c9f-fac8-4794-a69c-02178767e5b1") : configmap "swift-ring-files" not found Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.205860 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63b71752-f06a-4d1a-9d80-85142bbc3610","Type":"ContainerStarted","Data":"0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e"} Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.206344 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.207773 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.207806 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qchx9" event={"ID":"b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9","Type":"ContainerDied","Data":"67b911ea1d230ae5d572c17a6c603048f43d2af69ca2e20d0e7c9fd98fee5581"} Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.207833 4817 scope.go:117] "RemoveContainer" containerID="16e17218abb17dc509615f4bbde4f575898ac43de3b6626ce6201c990d1d7bcb" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.229445 4817 scope.go:117] "RemoveContainer" containerID="afafef7f7b2e467ef846e03f9647812e13d1f44a510620cc4cb229ad63e8762d" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.239280 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.033733842 podStartE2EDuration="1m4.23926567s" podCreationTimestamp="2025-11-28 14:44:45 +0000 UTC" firstStartedPulling="2025-11-28 14:44:47.093212966 +0000 UTC m=+909.681191232" lastFinishedPulling="2025-11-28 14:45:14.298744794 +0000 UTC m=+936.886723060" observedRunningTime="2025-11-28 14:45:49.235518537 +0000 UTC m=+971.823496803" watchObservedRunningTime="2025-11-28 14:45:49.23926567 +0000 UTC m=+971.827243936" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.276339 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qchx9"] Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.298973 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qchx9"] Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.706873 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-93e8-account-create-update-g8fpv"] Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707258 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de0b527-c06b-4250-854b-55c361288d1d" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707277 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de0b527-c06b-4250-854b-55c361288d1d" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707296 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707305 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707315 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="extract-utilities" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707323 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="extract-utilities" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707334 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707342 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707358 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707367 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707383 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="registry-server" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707390 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="registry-server" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707412 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd260782-cb7b-446c-8222-e28050effe0b" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707419 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd260782-cb7b-446c-8222-e28050effe0b" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707430 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707436 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707445 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd260782-cb7b-446c-8222-e28050effe0b" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707452 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd260782-cb7b-446c-8222-e28050effe0b" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: E1128 14:45:49.707462 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="extract-content" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707468 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="extract-content" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707661 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707683 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de0b527-c06b-4250-854b-55c361288d1d" containerName="init" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707698 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd260782-cb7b-446c-8222-e28050effe0b" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707711 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="34bea336-efd1-4432-8ccf-afc87bee4381" containerName="dnsmasq-dns" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.707741 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d3ce03-7abf-4cd7-8394-192f8c56ba51" containerName="registry-server" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.708367 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.710400 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.721323 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-93e8-account-create-update-g8fpv"] Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.753986 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9" path="/var/lib/kubelet/pods/b3ae3a23-d27a-4f7a-ac47-1cd2ed9439a9/volumes" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.761782 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-sdjgh"] Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.762804 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.778329 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sdjgh"] Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.852803 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-operator-scripts\") pod \"keystone-db-create-sdjgh\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.852859 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50fef50-1b7c-4f91-8f53-258516e099d7-operator-scripts\") pod \"keystone-93e8-account-create-update-g8fpv\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.852937 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czqr4\" (UniqueName: \"kubernetes.io/projected/b50fef50-1b7c-4f91-8f53-258516e099d7-kube-api-access-czqr4\") pod \"keystone-93e8-account-create-update-g8fpv\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.853011 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4vx8\" (UniqueName: \"kubernetes.io/projected/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-kube-api-access-m4vx8\") pod \"keystone-db-create-sdjgh\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.954183 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czqr4\" (UniqueName: \"kubernetes.io/projected/b50fef50-1b7c-4f91-8f53-258516e099d7-kube-api-access-czqr4\") pod \"keystone-93e8-account-create-update-g8fpv\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.954251 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4vx8\" (UniqueName: \"kubernetes.io/projected/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-kube-api-access-m4vx8\") pod \"keystone-db-create-sdjgh\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.954355 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-operator-scripts\") pod \"keystone-db-create-sdjgh\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.954386 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50fef50-1b7c-4f91-8f53-258516e099d7-operator-scripts\") pod \"keystone-93e8-account-create-update-g8fpv\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.955149 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50fef50-1b7c-4f91-8f53-258516e099d7-operator-scripts\") pod \"keystone-93e8-account-create-update-g8fpv\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.956015 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-operator-scripts\") pod \"keystone-db-create-sdjgh\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.959090 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-q4zm8"] Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.960059 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:49 crc kubenswrapper[4817]: I1128 14:45:49.977455 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q4zm8"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.006694 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.007106 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.011129 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4vx8\" (UniqueName: \"kubernetes.io/projected/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-kube-api-access-m4vx8\") pod \"keystone-db-create-sdjgh\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.013390 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czqr4\" (UniqueName: \"kubernetes.io/projected/b50fef50-1b7c-4f91-8f53-258516e099d7-kube-api-access-czqr4\") pod \"keystone-93e8-account-create-update-g8fpv\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.041217 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.044385 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c35b-account-create-update-jwxrp"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.045709 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.048433 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.054210 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c35b-account-create-update-jwxrp"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.059753 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbbxj\" (UniqueName: \"kubernetes.io/projected/92c6cb42-23bc-48e1-ba8a-671f454d5aac-kube-api-access-bbbxj\") pod \"placement-db-create-q4zm8\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.059812 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92c6cb42-23bc-48e1-ba8a-671f454d5aac-operator-scripts\") pod \"placement-db-create-q4zm8\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.080742 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.086046 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.161156 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-operator-scripts\") pod \"placement-c35b-account-create-update-jwxrp\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.161250 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbbxj\" (UniqueName: \"kubernetes.io/projected/92c6cb42-23bc-48e1-ba8a-671f454d5aac-kube-api-access-bbbxj\") pod \"placement-db-create-q4zm8\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.161314 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92c6cb42-23bc-48e1-ba8a-671f454d5aac-operator-scripts\") pod \"placement-db-create-q4zm8\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.161402 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd6tg\" (UniqueName: \"kubernetes.io/projected/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-kube-api-access-jd6tg\") pod \"placement-c35b-account-create-update-jwxrp\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.163246 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92c6cb42-23bc-48e1-ba8a-671f454d5aac-operator-scripts\") pod \"placement-db-create-q4zm8\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.188244 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbbxj\" (UniqueName: \"kubernetes.io/projected/92c6cb42-23bc-48e1-ba8a-671f454d5aac-kube-api-access-bbbxj\") pod \"placement-db-create-q4zm8\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.221127 4817 generic.go:334] "Generic (PLEG): container finished" podID="f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" containerID="801afa80588778b2ac91cca24026c967004c701c8409f822591f84050b871051" exitCode=0 Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.221174 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fzhb6" event={"ID":"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662","Type":"ContainerDied","Data":"801afa80588778b2ac91cca24026c967004c701c8409f822591f84050b871051"} Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.227855 4817 generic.go:334] "Generic (PLEG): container finished" podID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerID="8ccf30fb29efa6e931504451a614334aecdfe48ab9668d3b04ee44037bd78dc6" exitCode=0 Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.228630 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d340d5c2-67ed-48da-826f-e1e4c244832b","Type":"ContainerDied","Data":"8ccf30fb29efa6e931504451a614334aecdfe48ab9668d3b04ee44037bd78dc6"} Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.263182 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd6tg\" (UniqueName: \"kubernetes.io/projected/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-kube-api-access-jd6tg\") pod \"placement-c35b-account-create-update-jwxrp\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.267253 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-operator-scripts\") pod \"placement-c35b-account-create-update-jwxrp\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.268152 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-operator-scripts\") pod \"placement-c35b-account-create-update-jwxrp\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.286479 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.293091 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd6tg\" (UniqueName: \"kubernetes.io/projected/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-kube-api-access-jd6tg\") pod \"placement-c35b-account-create-update-jwxrp\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.295313 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.319106 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pvwxr"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.320161 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.330640 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pvwxr"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.363347 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bps4t"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.383870 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-93e8-account-create-update-g8fpv"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.423114 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-bf97-account-create-update-wb4z8"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.424233 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.428683 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.436136 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bf97-account-create-update-wb4z8"] Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.459452 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sdjgh"] Nov 28 14:45:50 crc kubenswrapper[4817]: W1128 14:45:50.469299 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod632a370a_92b6_4f1c_beeb_c6a7a5f562ff.slice/crio-4de271859f7ad5345976282d7c1b681fd069be9aa2207882bbc3726624200c75 WatchSource:0}: Error finding container 4de271859f7ad5345976282d7c1b681fd069be9aa2207882bbc3726624200c75: Status 404 returned error can't find the container with id 4de271859f7ad5345976282d7c1b681fd069be9aa2207882bbc3726624200c75 Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.472914 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r266r\" (UniqueName: \"kubernetes.io/projected/6c8e9ae0-c83a-4280-9d20-c03f045203d3-kube-api-access-r266r\") pod \"glance-bf97-account-create-update-wb4z8\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.472958 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c8e9ae0-c83a-4280-9d20-c03f045203d3-operator-scripts\") pod \"glance-bf97-account-create-update-wb4z8\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.473028 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-operator-scripts\") pod \"glance-db-create-pvwxr\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.473124 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7425\" (UniqueName: \"kubernetes.io/projected/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-kube-api-access-l7425\") pod \"glance-db-create-pvwxr\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.540575 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.574042 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7425\" (UniqueName: \"kubernetes.io/projected/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-kube-api-access-l7425\") pod \"glance-db-create-pvwxr\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.574134 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r266r\" (UniqueName: \"kubernetes.io/projected/6c8e9ae0-c83a-4280-9d20-c03f045203d3-kube-api-access-r266r\") pod \"glance-bf97-account-create-update-wb4z8\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.574152 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c8e9ae0-c83a-4280-9d20-c03f045203d3-operator-scripts\") pod \"glance-bf97-account-create-update-wb4z8\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.574208 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-operator-scripts\") pod \"glance-db-create-pvwxr\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.575275 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-operator-scripts\") pod \"glance-db-create-pvwxr\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.576515 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c8e9ae0-c83a-4280-9d20-c03f045203d3-operator-scripts\") pod \"glance-bf97-account-create-update-wb4z8\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.593128 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r266r\" (UniqueName: \"kubernetes.io/projected/6c8e9ae0-c83a-4280-9d20-c03f045203d3-kube-api-access-r266r\") pod \"glance-bf97-account-create-update-wb4z8\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.593997 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7425\" (UniqueName: \"kubernetes.io/projected/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-kube-api-access-l7425\") pod \"glance-db-create-pvwxr\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.649121 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.755955 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:50 crc kubenswrapper[4817]: I1128 14:45:50.793170 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q4zm8"] Nov 28 14:45:50 crc kubenswrapper[4817]: W1128 14:45:50.807805 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92c6cb42_23bc_48e1_ba8a_671f454d5aac.slice/crio-31af1d82aa64925cb59d6dca25e8f0e743b03c8d8a0043d0c17ec19e0e66af14 WatchSource:0}: Error finding container 31af1d82aa64925cb59d6dca25e8f0e743b03c8d8a0043d0c17ec19e0e66af14: Status 404 returned error can't find the container with id 31af1d82aa64925cb59d6dca25e8f0e743b03c8d8a0043d0c17ec19e0e66af14 Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.004313 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c35b-account-create-update-jwxrp"] Nov 28 14:45:51 crc kubenswrapper[4817]: W1128 14:45:51.029115 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c60d2ea_9bc6_48d2_a82a_788bf45b2381.slice/crio-3954ea84a6a44621e34c80a930d6aa0796ae1f5f474d07e756c3dc9b2afc7164 WatchSource:0}: Error finding container 3954ea84a6a44621e34c80a930d6aa0796ae1f5f474d07e756c3dc9b2afc7164: Status 404 returned error can't find the container with id 3954ea84a6a44621e34c80a930d6aa0796ae1f5f474d07e756c3dc9b2afc7164 Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.135869 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pvwxr"] Nov 28 14:45:51 crc kubenswrapper[4817]: W1128 14:45:51.155668 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeacbde1d_ce9d_460f_a9d4_e6bd7ea19c14.slice/crio-c96dd66f4c4a6744a248c34fd1bd60cf09ce1cadb143baefed7a0cc6305af84d WatchSource:0}: Error finding container c96dd66f4c4a6744a248c34fd1bd60cf09ce1cadb143baefed7a0cc6305af84d: Status 404 returned error can't find the container with id c96dd66f4c4a6744a248c34fd1bd60cf09ce1cadb143baefed7a0cc6305af84d Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.156286 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.239580 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4zm8" event={"ID":"92c6cb42-23bc-48e1-ba8a-671f454d5aac","Type":"ContainerStarted","Data":"e563b7daf40fd96fc19785df2b1cf6c5fc535eca965c42a1c00f6e9cc00b4ce5"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.239629 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4zm8" event={"ID":"92c6cb42-23bc-48e1-ba8a-671f454d5aac","Type":"ContainerStarted","Data":"31af1d82aa64925cb59d6dca25e8f0e743b03c8d8a0043d0c17ec19e0e66af14"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.244434 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvwxr" event={"ID":"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14","Type":"ContainerStarted","Data":"c96dd66f4c4a6744a248c34fd1bd60cf09ce1cadb143baefed7a0cc6305af84d"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.248921 4817 generic.go:334] "Generic (PLEG): container finished" podID="632a370a-92b6-4f1c-beeb-c6a7a5f562ff" containerID="d2006d514f24f53332336ad7fae97618d443db30e7155b77197a16537c18dba0" exitCode=0 Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.248989 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sdjgh" event={"ID":"632a370a-92b6-4f1c-beeb-c6a7a5f562ff","Type":"ContainerDied","Data":"d2006d514f24f53332336ad7fae97618d443db30e7155b77197a16537c18dba0"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.249038 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sdjgh" event={"ID":"632a370a-92b6-4f1c-beeb-c6a7a5f562ff","Type":"ContainerStarted","Data":"4de271859f7ad5345976282d7c1b681fd069be9aa2207882bbc3726624200c75"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.253114 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d340d5c2-67ed-48da-826f-e1e4c244832b","Type":"ContainerStarted","Data":"ed4c3b1acfccbff43f85441f72bfebd43f8cba89beaf216be01366e988d4c343"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.253896 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.257567 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-q4zm8" podStartSLOduration=2.25755197 podStartE2EDuration="2.25755197s" podCreationTimestamp="2025-11-28 14:45:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:45:51.254182186 +0000 UTC m=+973.842160452" watchObservedRunningTime="2025-11-28 14:45:51.25755197 +0000 UTC m=+973.845530236" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.267031 4817 generic.go:334] "Generic (PLEG): container finished" podID="b50fef50-1b7c-4f91-8f53-258516e099d7" containerID="4d985f2a6596a8516cc62873eeb632b2a30cfc58126fbd0d3ec15082ccc1d9bb" exitCode=0 Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.267122 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-93e8-account-create-update-g8fpv" event={"ID":"b50fef50-1b7c-4f91-8f53-258516e099d7","Type":"ContainerDied","Data":"4d985f2a6596a8516cc62873eeb632b2a30cfc58126fbd0d3ec15082ccc1d9bb"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.267162 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-93e8-account-create-update-g8fpv" event={"ID":"b50fef50-1b7c-4f91-8f53-258516e099d7","Type":"ContainerStarted","Data":"b4a6d5ac89ab506a126bd46dd6371d7b1aa348fa7c8ca4c1362a5511e533b6d1"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.269909 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c35b-account-create-update-jwxrp" event={"ID":"5c60d2ea-9bc6-48d2-a82a-788bf45b2381","Type":"ContainerStarted","Data":"3954ea84a6a44621e34c80a930d6aa0796ae1f5f474d07e756c3dc9b2afc7164"} Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.280668 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bf97-account-create-update-wb4z8"] Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.299751 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.378792462 podStartE2EDuration="1m6.29973451s" podCreationTimestamp="2025-11-28 14:44:45 +0000 UTC" firstStartedPulling="2025-11-28 14:44:57.451229946 +0000 UTC m=+920.039208222" lastFinishedPulling="2025-11-28 14:45:14.372172004 +0000 UTC m=+936.960150270" observedRunningTime="2025-11-28 14:45:51.29651919 +0000 UTC m=+973.884497446" watchObservedRunningTime="2025-11-28 14:45:51.29973451 +0000 UTC m=+973.887712776" Nov 28 14:45:51 crc kubenswrapper[4817]: E1128 14:45:51.536283 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92c6cb42_23bc_48e1_ba8a_671f454d5aac.slice/crio-e563b7daf40fd96fc19785df2b1cf6c5fc535eca965c42a1c00f6e9cc00b4ce5.scope\": RecentStats: unable to find data in memory cache]" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.591567 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700450 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-dispersionconf\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700515 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-ring-data-devices\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700544 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-scripts\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700646 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khlcc\" (UniqueName: \"kubernetes.io/projected/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-kube-api-access-khlcc\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700732 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-combined-ca-bundle\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700779 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-etc-swift\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.700799 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-swiftconf\") pod \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\" (UID: \"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662\") " Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.701019 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.701462 4817 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.701877 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.707410 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-kube-api-access-khlcc" (OuterVolumeSpecName: "kube-api-access-khlcc") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "kube-api-access-khlcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.712855 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.726853 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.728253 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.729192 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-scripts" (OuterVolumeSpecName: "scripts") pod "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" (UID: "f61aa2a2-1ba8-4319-ad02-2c05d6a9f662"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.804137 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.804168 4817 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.804182 4817 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.804190 4817 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.804198 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:51 crc kubenswrapper[4817]: I1128 14:45:51.804206 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khlcc\" (UniqueName: \"kubernetes.io/projected/f61aa2a2-1ba8-4319-ad02-2c05d6a9f662-kube-api-access-khlcc\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.279065 4817 generic.go:334] "Generic (PLEG): container finished" podID="5c60d2ea-9bc6-48d2-a82a-788bf45b2381" containerID="599f2eb3493ecf17c978bbf915382f7a92c014240322321f14cb98433c849039" exitCode=0 Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.279114 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c35b-account-create-update-jwxrp" event={"ID":"5c60d2ea-9bc6-48d2-a82a-788bf45b2381","Type":"ContainerDied","Data":"599f2eb3493ecf17c978bbf915382f7a92c014240322321f14cb98433c849039"} Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.281322 4817 generic.go:334] "Generic (PLEG): container finished" podID="92c6cb42-23bc-48e1-ba8a-671f454d5aac" containerID="e563b7daf40fd96fc19785df2b1cf6c5fc535eca965c42a1c00f6e9cc00b4ce5" exitCode=0 Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.281362 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4zm8" event={"ID":"92c6cb42-23bc-48e1-ba8a-671f454d5aac","Type":"ContainerDied","Data":"e563b7daf40fd96fc19785df2b1cf6c5fc535eca965c42a1c00f6e9cc00b4ce5"} Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.283249 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fzhb6" event={"ID":"f61aa2a2-1ba8-4319-ad02-2c05d6a9f662","Type":"ContainerDied","Data":"1cd7822c724de04b6cc4dcf022fa2dc75926152d577d7478166b8c59e0828d3d"} Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.283276 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cd7822c724de04b6cc4dcf022fa2dc75926152d577d7478166b8c59e0828d3d" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.283342 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fzhb6" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.285139 4817 generic.go:334] "Generic (PLEG): container finished" podID="6c8e9ae0-c83a-4280-9d20-c03f045203d3" containerID="8dadfdf37a4c2d9a4084785cf7d5fab3e8860e24e144650fa6cd148c8d468748" exitCode=0 Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.285217 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bf97-account-create-update-wb4z8" event={"ID":"6c8e9ae0-c83a-4280-9d20-c03f045203d3","Type":"ContainerDied","Data":"8dadfdf37a4c2d9a4084785cf7d5fab3e8860e24e144650fa6cd148c8d468748"} Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.285248 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bf97-account-create-update-wb4z8" event={"ID":"6c8e9ae0-c83a-4280-9d20-c03f045203d3","Type":"ContainerStarted","Data":"63756a797beb8493bd76df5ca3efeee1771eecaaf539dbd2252d406aa768c1c8"} Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.287598 4817 generic.go:334] "Generic (PLEG): container finished" podID="eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" containerID="ba50ebff34de984713a13742308c1ad582ca9e8dbbc1eb26869ea57d72d815da" exitCode=0 Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.288496 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvwxr" event={"ID":"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14","Type":"ContainerDied","Data":"ba50ebff34de984713a13742308c1ad582ca9e8dbbc1eb26869ea57d72d815da"} Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.288624 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bps4t" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="registry-server" containerID="cri-o://9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb" gracePeriod=2 Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.705569 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.815109 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.819899 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50fef50-1b7c-4f91-8f53-258516e099d7-operator-scripts\") pod \"b50fef50-1b7c-4f91-8f53-258516e099d7\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.820105 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czqr4\" (UniqueName: \"kubernetes.io/projected/b50fef50-1b7c-4f91-8f53-258516e099d7-kube-api-access-czqr4\") pod \"b50fef50-1b7c-4f91-8f53-258516e099d7\" (UID: \"b50fef50-1b7c-4f91-8f53-258516e099d7\") " Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.820364 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50fef50-1b7c-4f91-8f53-258516e099d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b50fef50-1b7c-4f91-8f53-258516e099d7" (UID: "b50fef50-1b7c-4f91-8f53-258516e099d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.820491 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50fef50-1b7c-4f91-8f53-258516e099d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.824801 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b50fef50-1b7c-4f91-8f53-258516e099d7-kube-api-access-czqr4" (OuterVolumeSpecName: "kube-api-access-czqr4") pod "b50fef50-1b7c-4f91-8f53-258516e099d7" (UID: "b50fef50-1b7c-4f91-8f53-258516e099d7"). InnerVolumeSpecName "kube-api-access-czqr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.921419 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4vx8\" (UniqueName: \"kubernetes.io/projected/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-kube-api-access-m4vx8\") pod \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.921509 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-operator-scripts\") pod \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\" (UID: \"632a370a-92b6-4f1c-beeb-c6a7a5f562ff\") " Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.921865 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czqr4\" (UniqueName: \"kubernetes.io/projected/b50fef50-1b7c-4f91-8f53-258516e099d7-kube-api-access-czqr4\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.921965 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "632a370a-92b6-4f1c-beeb-c6a7a5f562ff" (UID: "632a370a-92b6-4f1c-beeb-c6a7a5f562ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.924713 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-kube-api-access-m4vx8" (OuterVolumeSpecName: "kube-api-access-m4vx8") pod "632a370a-92b6-4f1c-beeb-c6a7a5f562ff" (UID: "632a370a-92b6-4f1c-beeb-c6a7a5f562ff"). InnerVolumeSpecName "kube-api-access-m4vx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:52 crc kubenswrapper[4817]: I1128 14:45:52.928792 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.023169 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-catalog-content\") pod \"43d670c6-8737-40ec-9a71-f8f7347a0c97\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.023341 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mfgr\" (UniqueName: \"kubernetes.io/projected/43d670c6-8737-40ec-9a71-f8f7347a0c97-kube-api-access-7mfgr\") pod \"43d670c6-8737-40ec-9a71-f8f7347a0c97\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.023436 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-utilities\") pod \"43d670c6-8737-40ec-9a71-f8f7347a0c97\" (UID: \"43d670c6-8737-40ec-9a71-f8f7347a0c97\") " Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.024018 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4vx8\" (UniqueName: \"kubernetes.io/projected/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-kube-api-access-m4vx8\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.024038 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/632a370a-92b6-4f1c-beeb-c6a7a5f562ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.024258 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-utilities" (OuterVolumeSpecName: "utilities") pod "43d670c6-8737-40ec-9a71-f8f7347a0c97" (UID: "43d670c6-8737-40ec-9a71-f8f7347a0c97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.026858 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43d670c6-8737-40ec-9a71-f8f7347a0c97-kube-api-access-7mfgr" (OuterVolumeSpecName: "kube-api-access-7mfgr") pod "43d670c6-8737-40ec-9a71-f8f7347a0c97" (UID: "43d670c6-8737-40ec-9a71-f8f7347a0c97"). InnerVolumeSpecName "kube-api-access-7mfgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.045251 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43d670c6-8737-40ec-9a71-f8f7347a0c97" (UID: "43d670c6-8737-40ec-9a71-f8f7347a0c97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.125908 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mfgr\" (UniqueName: \"kubernetes.io/projected/43d670c6-8737-40ec-9a71-f8f7347a0c97-kube-api-access-7mfgr\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.125947 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.125959 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43d670c6-8737-40ec-9a71-f8f7347a0c97-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.295379 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sdjgh" event={"ID":"632a370a-92b6-4f1c-beeb-c6a7a5f562ff","Type":"ContainerDied","Data":"4de271859f7ad5345976282d7c1b681fd069be9aa2207882bbc3726624200c75"} Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.295436 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4de271859f7ad5345976282d7c1b681fd069be9aa2207882bbc3726624200c75" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.295383 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sdjgh" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.296902 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-93e8-account-create-update-g8fpv" event={"ID":"b50fef50-1b7c-4f91-8f53-258516e099d7","Type":"ContainerDied","Data":"b4a6d5ac89ab506a126bd46dd6371d7b1aa348fa7c8ca4c1362a5511e533b6d1"} Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.296947 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4a6d5ac89ab506a126bd46dd6371d7b1aa348fa7c8ca4c1362a5511e533b6d1" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.296958 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-93e8-account-create-update-g8fpv" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.298586 4817 generic.go:334] "Generic (PLEG): container finished" podID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerID="9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb" exitCode=0 Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.298633 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bps4t" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.298683 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bps4t" event={"ID":"43d670c6-8737-40ec-9a71-f8f7347a0c97","Type":"ContainerDied","Data":"9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb"} Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.298706 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bps4t" event={"ID":"43d670c6-8737-40ec-9a71-f8f7347a0c97","Type":"ContainerDied","Data":"f492db1a2c5aface0ddab77845cb54b022d6a0b4320d71c55a659fe0fc2487c2"} Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.298741 4817 scope.go:117] "RemoveContainer" containerID="9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.355615 4817 scope.go:117] "RemoveContainer" containerID="997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.359761 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bps4t"] Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.367401 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bps4t"] Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.398315 4817 scope.go:117] "RemoveContainer" containerID="facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.460635 4817 scope.go:117] "RemoveContainer" containerID="9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb" Nov 28 14:45:53 crc kubenswrapper[4817]: E1128 14:45:53.463029 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb\": container with ID starting with 9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb not found: ID does not exist" containerID="9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.463301 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb"} err="failed to get container status \"9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb\": rpc error: code = NotFound desc = could not find container \"9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb\": container with ID starting with 9070206b735497e565683f251ff7e13382d8dab7e094b522bcfeb0ef411b81fb not found: ID does not exist" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.463421 4817 scope.go:117] "RemoveContainer" containerID="997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756" Nov 28 14:45:53 crc kubenswrapper[4817]: E1128 14:45:53.463953 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756\": container with ID starting with 997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756 not found: ID does not exist" containerID="997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.464007 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756"} err="failed to get container status \"997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756\": rpc error: code = NotFound desc = could not find container \"997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756\": container with ID starting with 997314e18b6595472ca67ee0cebc6596a3c72ad7e8e0c3789d3a9cd85852f756 not found: ID does not exist" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.464039 4817 scope.go:117] "RemoveContainer" containerID="facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d" Nov 28 14:45:53 crc kubenswrapper[4817]: E1128 14:45:53.464463 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d\": container with ID starting with facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d not found: ID does not exist" containerID="facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.464586 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d"} err="failed to get container status \"facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d\": rpc error: code = NotFound desc = could not find container \"facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d\": container with ID starting with facbc68d7e080f82d677376eb1610e8df5f77aec130aba2fca0161052400be8d not found: ID does not exist" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.661062 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.747060 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" path="/var/lib/kubelet/pods/43d670c6-8737-40ec-9a71-f8f7347a0c97/volumes" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.836769 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbbxj\" (UniqueName: \"kubernetes.io/projected/92c6cb42-23bc-48e1-ba8a-671f454d5aac-kube-api-access-bbbxj\") pod \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.836918 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92c6cb42-23bc-48e1-ba8a-671f454d5aac-operator-scripts\") pod \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\" (UID: \"92c6cb42-23bc-48e1-ba8a-671f454d5aac\") " Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.838679 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c6cb42-23bc-48e1-ba8a-671f454d5aac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92c6cb42-23bc-48e1-ba8a-671f454d5aac" (UID: "92c6cb42-23bc-48e1-ba8a-671f454d5aac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.845377 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c6cb42-23bc-48e1-ba8a-671f454d5aac-kube-api-access-bbbxj" (OuterVolumeSpecName: "kube-api-access-bbbxj") pod "92c6cb42-23bc-48e1-ba8a-671f454d5aac" (UID: "92c6cb42-23bc-48e1-ba8a-671f454d5aac"). InnerVolumeSpecName "kube-api-access-bbbxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.866828 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.880118 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.901088 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.938339 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92c6cb42-23bc-48e1-ba8a-671f454d5aac-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:53 crc kubenswrapper[4817]: I1128 14:45:53.938375 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbbxj\" (UniqueName: \"kubernetes.io/projected/92c6cb42-23bc-48e1-ba8a-671f454d5aac-kube-api-access-bbbxj\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.039282 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r266r\" (UniqueName: \"kubernetes.io/projected/6c8e9ae0-c83a-4280-9d20-c03f045203d3-kube-api-access-r266r\") pod \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.039351 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-operator-scripts\") pod \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.039406 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7425\" (UniqueName: \"kubernetes.io/projected/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-kube-api-access-l7425\") pod \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.039431 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd6tg\" (UniqueName: \"kubernetes.io/projected/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-kube-api-access-jd6tg\") pod \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\" (UID: \"5c60d2ea-9bc6-48d2-a82a-788bf45b2381\") " Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.039642 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-operator-scripts\") pod \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\" (UID: \"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14\") " Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.039783 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c8e9ae0-c83a-4280-9d20-c03f045203d3-operator-scripts\") pod \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\" (UID: \"6c8e9ae0-c83a-4280-9d20-c03f045203d3\") " Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.040823 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c60d2ea-9bc6-48d2-a82a-788bf45b2381" (UID: "5c60d2ea-9bc6-48d2-a82a-788bf45b2381"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.040929 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c8e9ae0-c83a-4280-9d20-c03f045203d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c8e9ae0-c83a-4280-9d20-c03f045203d3" (UID: "6c8e9ae0-c83a-4280-9d20-c03f045203d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.040872 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" (UID: "eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.041167 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.041197 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.044006 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c8e9ae0-c83a-4280-9d20-c03f045203d3-kube-api-access-r266r" (OuterVolumeSpecName: "kube-api-access-r266r") pod "6c8e9ae0-c83a-4280-9d20-c03f045203d3" (UID: "6c8e9ae0-c83a-4280-9d20-c03f045203d3"). InnerVolumeSpecName "kube-api-access-r266r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.045390 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.045444 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.046016 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-kube-api-access-l7425" (OuterVolumeSpecName: "kube-api-access-l7425") pod "eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" (UID: "eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14"). InnerVolumeSpecName "kube-api-access-l7425". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.047422 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-kube-api-access-jd6tg" (OuterVolumeSpecName: "kube-api-access-jd6tg") pod "5c60d2ea-9bc6-48d2-a82a-788bf45b2381" (UID: "5c60d2ea-9bc6-48d2-a82a-788bf45b2381"). InnerVolumeSpecName "kube-api-access-jd6tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.142735 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r266r\" (UniqueName: \"kubernetes.io/projected/6c8e9ae0-c83a-4280-9d20-c03f045203d3-kube-api-access-r266r\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.142797 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7425\" (UniqueName: \"kubernetes.io/projected/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14-kube-api-access-l7425\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.142812 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd6tg\" (UniqueName: \"kubernetes.io/projected/5c60d2ea-9bc6-48d2-a82a-788bf45b2381-kube-api-access-jd6tg\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.142825 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c8e9ae0-c83a-4280-9d20-c03f045203d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.307909 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q4zm8" event={"ID":"92c6cb42-23bc-48e1-ba8a-671f454d5aac","Type":"ContainerDied","Data":"31af1d82aa64925cb59d6dca25e8f0e743b03c8d8a0043d0c17ec19e0e66af14"} Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.307934 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q4zm8" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.307947 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31af1d82aa64925cb59d6dca25e8f0e743b03c8d8a0043d0c17ec19e0e66af14" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.309400 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bf97-account-create-update-wb4z8" event={"ID":"6c8e9ae0-c83a-4280-9d20-c03f045203d3","Type":"ContainerDied","Data":"63756a797beb8493bd76df5ca3efeee1771eecaaf539dbd2252d406aa768c1c8"} Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.309426 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63756a797beb8493bd76df5ca3efeee1771eecaaf539dbd2252d406aa768c1c8" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.309426 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bf97-account-create-update-wb4z8" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.310974 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pvwxr" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.311057 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pvwxr" event={"ID":"eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14","Type":"ContainerDied","Data":"c96dd66f4c4a6744a248c34fd1bd60cf09ce1cadb143baefed7a0cc6305af84d"} Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.311151 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c96dd66f4c4a6744a248c34fd1bd60cf09ce1cadb143baefed7a0cc6305af84d" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.312194 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c35b-account-create-update-jwxrp" event={"ID":"5c60d2ea-9bc6-48d2-a82a-788bf45b2381","Type":"ContainerDied","Data":"3954ea84a6a44621e34c80a930d6aa0796ae1f5f474d07e756c3dc9b2afc7164"} Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.312281 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3954ea84a6a44621e34c80a930d6aa0796ae1f5f474d07e756c3dc9b2afc7164" Nov 28 14:45:54 crc kubenswrapper[4817]: I1128 14:45:54.312386 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c35b-account-create-update-jwxrp" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.541151 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-5gzqs"] Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.542949 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c60d2ea-9bc6-48d2-a82a-788bf45b2381" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.543054 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c60d2ea-9bc6-48d2-a82a-788bf45b2381" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.543172 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c6cb42-23bc-48e1-ba8a-671f454d5aac" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.543239 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c6cb42-23bc-48e1-ba8a-671f454d5aac" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.543329 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" containerName="swift-ring-rebalance" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.543412 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" containerName="swift-ring-rebalance" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.543503 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="extract-utilities" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.543570 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="extract-utilities" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.543648 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8e9ae0-c83a-4280-9d20-c03f045203d3" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.543745 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8e9ae0-c83a-4280-9d20-c03f045203d3" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.543852 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632a370a-92b6-4f1c-beeb-c6a7a5f562ff" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.543926 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="632a370a-92b6-4f1c-beeb-c6a7a5f562ff" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.544004 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50fef50-1b7c-4f91-8f53-258516e099d7" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.544080 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50fef50-1b7c-4f91-8f53-258516e099d7" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.544160 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.544235 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.544309 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="extract-content" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.544388 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="extract-content" Nov 28 14:45:55 crc kubenswrapper[4817]: E1128 14:45:55.544486 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="registry-server" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.544546 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="registry-server" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545000 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545109 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b50fef50-1b7c-4f91-8f53-258516e099d7" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545206 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c6cb42-23bc-48e1-ba8a-671f454d5aac" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545282 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="632a370a-92b6-4f1c-beeb-c6a7a5f562ff" containerName="mariadb-database-create" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545351 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f61aa2a2-1ba8-4319-ad02-2c05d6a9f662" containerName="swift-ring-rebalance" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545420 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8e9ae0-c83a-4280-9d20-c03f045203d3" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545503 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c60d2ea-9bc6-48d2-a82a-788bf45b2381" containerName="mariadb-account-create-update" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.545577 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d670c6-8737-40ec-9a71-f8f7347a0c97" containerName="registry-server" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.546624 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.550972 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5gzqs"] Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.551759 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wtq97" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.558526 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.664453 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-config-data\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.664518 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-db-sync-config-data\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.664560 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-combined-ca-bundle\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.664901 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66s8r\" (UniqueName: \"kubernetes.io/projected/77135fc7-4af7-4119-9153-525f2e7862c4-kube-api-access-66s8r\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.707854 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-pwgz2" podUID="0f58e7f4-f831-4f54-91b6-a5e97bf5152c" containerName="ovn-controller" probeResult="failure" output=< Nov 28 14:45:55 crc kubenswrapper[4817]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 14:45:55 crc kubenswrapper[4817]: > Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.767716 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-combined-ca-bundle\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.767923 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66s8r\" (UniqueName: \"kubernetes.io/projected/77135fc7-4af7-4119-9153-525f2e7862c4-kube-api-access-66s8r\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.768046 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-config-data\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.768153 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-db-sync-config-data\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.777297 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-db-sync-config-data\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.784327 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-config-data\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.785561 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-combined-ca-bundle\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.790345 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66s8r\" (UniqueName: \"kubernetes.io/projected/77135fc7-4af7-4119-9153-525f2e7862c4-kube-api-access-66s8r\") pod \"glance-db-sync-5gzqs\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:55 crc kubenswrapper[4817]: I1128 14:45:55.868535 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5gzqs" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.089646 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.092900 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vz6mt" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.317414 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pwgz2-config-q8kgz"] Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.319308 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.327817 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.356052 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pwgz2-config-q8kgz"] Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.412279 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5gzqs"] Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.482034 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-log-ovn\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.482635 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-additional-scripts\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.482861 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run-ovn\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.483005 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-scripts\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.483140 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wvbm\" (UniqueName: \"kubernetes.io/projected/acd60d36-a514-4a6b-977b-95bd327dce6b-kube-api-access-5wvbm\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.483299 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.584614 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-log-ovn\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586009 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-log-ovn\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586007 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-additional-scripts\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586266 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run-ovn\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586381 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-scripts\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586466 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wvbm\" (UniqueName: \"kubernetes.io/projected/acd60d36-a514-4a6b-977b-95bd327dce6b-kube-api-access-5wvbm\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586567 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586678 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586420 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run-ovn\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.586763 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-additional-scripts\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.588448 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-scripts\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.618567 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wvbm\" (UniqueName: \"kubernetes.io/projected/acd60d36-a514-4a6b-977b-95bd327dce6b-kube-api-access-5wvbm\") pod \"ovn-controller-pwgz2-config-q8kgz\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:56 crc kubenswrapper[4817]: I1128 14:45:56.659913 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:57 crc kubenswrapper[4817]: I1128 14:45:57.069193 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pwgz2-config-q8kgz"] Nov 28 14:45:57 crc kubenswrapper[4817]: I1128 14:45:57.365900 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5gzqs" event={"ID":"77135fc7-4af7-4119-9153-525f2e7862c4","Type":"ContainerStarted","Data":"0107cc5b36e5fad4599d78db2823d09c2275d7a97c6cfa1863565f325fe8d786"} Nov 28 14:45:57 crc kubenswrapper[4817]: I1128 14:45:57.368491 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-q8kgz" event={"ID":"acd60d36-a514-4a6b-977b-95bd327dce6b","Type":"ContainerStarted","Data":"9e8c095cafb5dfbb3c775ef8f042420e9d9c421fd42257b5f4fc354ea195c940"} Nov 28 14:45:57 crc kubenswrapper[4817]: I1128 14:45:57.368548 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-q8kgz" event={"ID":"acd60d36-a514-4a6b-977b-95bd327dce6b","Type":"ContainerStarted","Data":"cf5604d0895b1de30a5ef99dbb2953aa7c062221e474f42a80627f281bcfc7d5"} Nov 28 14:45:57 crc kubenswrapper[4817]: I1128 14:45:57.763126 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pwgz2-config-q8kgz" podStartSLOduration=1.763105027 podStartE2EDuration="1.763105027s" podCreationTimestamp="2025-11-28 14:45:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:45:57.388481454 +0000 UTC m=+979.976459730" watchObservedRunningTime="2025-11-28 14:45:57.763105027 +0000 UTC m=+980.351083293" Nov 28 14:45:58 crc kubenswrapper[4817]: I1128 14:45:58.386250 4817 generic.go:334] "Generic (PLEG): container finished" podID="acd60d36-a514-4a6b-977b-95bd327dce6b" containerID="9e8c095cafb5dfbb3c775ef8f042420e9d9c421fd42257b5f4fc354ea195c940" exitCode=0 Nov 28 14:45:58 crc kubenswrapper[4817]: I1128 14:45:58.386517 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-q8kgz" event={"ID":"acd60d36-a514-4a6b-977b-95bd327dce6b","Type":"ContainerDied","Data":"9e8c095cafb5dfbb3c775ef8f042420e9d9c421fd42257b5f4fc354ea195c940"} Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.763200 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.939807 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-additional-scripts\") pod \"acd60d36-a514-4a6b-977b-95bd327dce6b\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940278 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run\") pod \"acd60d36-a514-4a6b-977b-95bd327dce6b\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940317 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-scripts\") pod \"acd60d36-a514-4a6b-977b-95bd327dce6b\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940394 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run-ovn\") pod \"acd60d36-a514-4a6b-977b-95bd327dce6b\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940433 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-log-ovn\") pod \"acd60d36-a514-4a6b-977b-95bd327dce6b\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940446 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run" (OuterVolumeSpecName: "var-run") pod "acd60d36-a514-4a6b-977b-95bd327dce6b" (UID: "acd60d36-a514-4a6b-977b-95bd327dce6b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940492 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wvbm\" (UniqueName: \"kubernetes.io/projected/acd60d36-a514-4a6b-977b-95bd327dce6b-kube-api-access-5wvbm\") pod \"acd60d36-a514-4a6b-977b-95bd327dce6b\" (UID: \"acd60d36-a514-4a6b-977b-95bd327dce6b\") " Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940766 4817 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.940928 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "acd60d36-a514-4a6b-977b-95bd327dce6b" (UID: "acd60d36-a514-4a6b-977b-95bd327dce6b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.941447 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "acd60d36-a514-4a6b-977b-95bd327dce6b" (UID: "acd60d36-a514-4a6b-977b-95bd327dce6b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.941525 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "acd60d36-a514-4a6b-977b-95bd327dce6b" (UID: "acd60d36-a514-4a6b-977b-95bd327dce6b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.941902 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-scripts" (OuterVolumeSpecName: "scripts") pod "acd60d36-a514-4a6b-977b-95bd327dce6b" (UID: "acd60d36-a514-4a6b-977b-95bd327dce6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:45:59 crc kubenswrapper[4817]: I1128 14:45:59.954434 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd60d36-a514-4a6b-977b-95bd327dce6b-kube-api-access-5wvbm" (OuterVolumeSpecName: "kube-api-access-5wvbm") pod "acd60d36-a514-4a6b-977b-95bd327dce6b" (UID: "acd60d36-a514-4a6b-977b-95bd327dce6b"). InnerVolumeSpecName "kube-api-access-5wvbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.041896 4817 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.041930 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acd60d36-a514-4a6b-977b-95bd327dce6b-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.041939 4817 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.041946 4817 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acd60d36-a514-4a6b-977b-95bd327dce6b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.041955 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wvbm\" (UniqueName: \"kubernetes.io/projected/acd60d36-a514-4a6b-977b-95bd327dce6b-kube-api-access-5wvbm\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.410103 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-q8kgz" event={"ID":"acd60d36-a514-4a6b-977b-95bd327dce6b","Type":"ContainerDied","Data":"cf5604d0895b1de30a5ef99dbb2953aa7c062221e474f42a80627f281bcfc7d5"} Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.410186 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf5604d0895b1de30a5ef99dbb2953aa7c062221e474f42a80627f281bcfc7d5" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.410193 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-q8kgz" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.488417 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pwgz2-config-q8kgz"] Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.503258 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pwgz2-config-q8kgz"] Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.605828 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pwgz2-config-shlhj"] Nov 28 14:46:00 crc kubenswrapper[4817]: E1128 14:46:00.606120 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd60d36-a514-4a6b-977b-95bd327dce6b" containerName="ovn-config" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.606136 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd60d36-a514-4a6b-977b-95bd327dce6b" containerName="ovn-config" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.606309 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd60d36-a514-4a6b-977b-95bd327dce6b" containerName="ovn-config" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.606798 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.609426 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.619106 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pwgz2-config-shlhj"] Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.700629 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-pwgz2" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.751982 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqjhz\" (UniqueName: \"kubernetes.io/projected/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-kube-api-access-cqjhz\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.752528 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run-ovn\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.752589 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-additional-scripts\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.752620 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-log-ovn\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.752733 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-scripts\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.752781 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.853850 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqjhz\" (UniqueName: \"kubernetes.io/projected/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-kube-api-access-cqjhz\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.853959 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run-ovn\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.853998 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-additional-scripts\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.854669 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-log-ovn\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.854690 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-scripts\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.854709 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.854747 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-log-ovn\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.854867 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.855111 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-additional-scripts\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.855685 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run-ovn\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.856627 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-scripts\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.876435 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqjhz\" (UniqueName: \"kubernetes.io/projected/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-kube-api-access-cqjhz\") pod \"ovn-controller-pwgz2-config-shlhj\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:00 crc kubenswrapper[4817]: I1128 14:46:00.922756 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:01 crc kubenswrapper[4817]: I1128 14:46:01.384954 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pwgz2-config-shlhj"] Nov 28 14:46:01 crc kubenswrapper[4817]: W1128 14:46:01.393694 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92e9f311_26cd_4c4b_9aa6_d3255ccb2d13.slice/crio-d84727f28322476525ecba480dc0e76d0bdff462019470251aa9cb70976ceed6 WatchSource:0}: Error finding container d84727f28322476525ecba480dc0e76d0bdff462019470251aa9cb70976ceed6: Status 404 returned error can't find the container with id d84727f28322476525ecba480dc0e76d0bdff462019470251aa9cb70976ceed6 Nov 28 14:46:01 crc kubenswrapper[4817]: I1128 14:46:01.418669 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-shlhj" event={"ID":"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13","Type":"ContainerStarted","Data":"d84727f28322476525ecba480dc0e76d0bdff462019470251aa9cb70976ceed6"} Nov 28 14:46:01 crc kubenswrapper[4817]: I1128 14:46:01.750204 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd60d36-a514-4a6b-977b-95bd327dce6b" path="/var/lib/kubelet/pods/acd60d36-a514-4a6b-977b-95bd327dce6b/volumes" Nov 28 14:46:02 crc kubenswrapper[4817]: I1128 14:46:02.432750 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-shlhj" event={"ID":"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13","Type":"ContainerStarted","Data":"e00dd23c0aaeb25c7291bfe0d3d52cc13f85f7a4254c3df7cfd0281423e2c128"} Nov 28 14:46:02 crc kubenswrapper[4817]: I1128 14:46:02.464097 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pwgz2-config-shlhj" podStartSLOduration=2.464073329 podStartE2EDuration="2.464073329s" podCreationTimestamp="2025-11-28 14:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:46:02.448258595 +0000 UTC m=+985.036236881" watchObservedRunningTime="2025-11-28 14:46:02.464073329 +0000 UTC m=+985.052051595" Nov 28 14:46:03 crc kubenswrapper[4817]: I1128 14:46:03.440613 4817 generic.go:334] "Generic (PLEG): container finished" podID="92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" containerID="e00dd23c0aaeb25c7291bfe0d3d52cc13f85f7a4254c3df7cfd0281423e2c128" exitCode=0 Nov 28 14:46:03 crc kubenswrapper[4817]: I1128 14:46:03.440655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-shlhj" event={"ID":"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13","Type":"ContainerDied","Data":"e00dd23c0aaeb25c7291bfe0d3d52cc13f85f7a4254c3df7cfd0281423e2c128"} Nov 28 14:46:05 crc kubenswrapper[4817]: I1128 14:46:05.146447 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:46:05 crc kubenswrapper[4817]: I1128 14:46:05.158668 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42276c9f-fac8-4794-a69c-02178767e5b1-etc-swift\") pod \"swift-storage-0\" (UID: \"42276c9f-fac8-4794-a69c-02178767e5b1\") " pod="openstack/swift-storage-0" Nov 28 14:46:05 crc kubenswrapper[4817]: I1128 14:46:05.402507 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 14:46:06 crc kubenswrapper[4817]: I1128 14:46:06.533959 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 28 14:46:06 crc kubenswrapper[4817]: I1128 14:46:06.837901 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.006646 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-5p8hr"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.008856 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.016896 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f9d9-account-create-update-v74l9"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.017945 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.020868 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.053268 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5p8hr"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.062794 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f9d9-account-create-update-v74l9"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.126877 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-pbdln"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.128058 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.134930 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pbdln"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.193918 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdrbn\" (UniqueName: \"kubernetes.io/projected/eb9853b7-9f8c-4b0c-8b02-4bf014636556-kube-api-access-mdrbn\") pod \"cinder-f9d9-account-create-update-v74l9\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.193978 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb9853b7-9f8c-4b0c-8b02-4bf014636556-operator-scripts\") pod \"cinder-f9d9-account-create-update-v74l9\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.194096 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56eb98d9-0aff-4113-b6f0-34329d3b476d-operator-scripts\") pod \"cinder-db-create-5p8hr\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.194369 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcq5k\" (UniqueName: \"kubernetes.io/projected/56eb98d9-0aff-4113-b6f0-34329d3b476d-kube-api-access-lcq5k\") pod \"cinder-db-create-5p8hr\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.260511 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gzplt"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.261889 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.264772 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.265178 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-knk6m" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.265532 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.265579 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.274537 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gzplt"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.296013 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56eb98d9-0aff-4113-b6f0-34329d3b476d-operator-scripts\") pod \"cinder-db-create-5p8hr\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.296082 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66v94\" (UniqueName: \"kubernetes.io/projected/7599e834-c07e-4db9-828b-675b53df2aa6-kube-api-access-66v94\") pod \"barbican-db-create-pbdln\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.296177 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7599e834-c07e-4db9-828b-675b53df2aa6-operator-scripts\") pod \"barbican-db-create-pbdln\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.296216 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcq5k\" (UniqueName: \"kubernetes.io/projected/56eb98d9-0aff-4113-b6f0-34329d3b476d-kube-api-access-lcq5k\") pod \"cinder-db-create-5p8hr\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.296340 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdrbn\" (UniqueName: \"kubernetes.io/projected/eb9853b7-9f8c-4b0c-8b02-4bf014636556-kube-api-access-mdrbn\") pod \"cinder-f9d9-account-create-update-v74l9\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.296379 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb9853b7-9f8c-4b0c-8b02-4bf014636556-operator-scripts\") pod \"cinder-f9d9-account-create-update-v74l9\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.297272 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb9853b7-9f8c-4b0c-8b02-4bf014636556-operator-scripts\") pod \"cinder-f9d9-account-create-update-v74l9\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.297892 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56eb98d9-0aff-4113-b6f0-34329d3b476d-operator-scripts\") pod \"cinder-db-create-5p8hr\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.307351 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-49a4-account-create-update-vmg9v"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.308340 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.315085 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.325390 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcq5k\" (UniqueName: \"kubernetes.io/projected/56eb98d9-0aff-4113-b6f0-34329d3b476d-kube-api-access-lcq5k\") pod \"cinder-db-create-5p8hr\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.327612 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-tncs8"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.328917 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.333760 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.338334 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdrbn\" (UniqueName: \"kubernetes.io/projected/eb9853b7-9f8c-4b0c-8b02-4bf014636556-kube-api-access-mdrbn\") pod \"cinder-f9d9-account-create-update-v74l9\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.347177 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.354680 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-49a4-account-create-update-vmg9v"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.361381 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-tncs8"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.397932 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66v94\" (UniqueName: \"kubernetes.io/projected/7599e834-c07e-4db9-828b-675b53df2aa6-kube-api-access-66v94\") pod \"barbican-db-create-pbdln\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.398013 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7599e834-c07e-4db9-828b-675b53df2aa6-operator-scripts\") pod \"barbican-db-create-pbdln\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.398046 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t8wd\" (UniqueName: \"kubernetes.io/projected/493a9cc0-bb5a-463c-be84-c6def176bac7-kube-api-access-2t8wd\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.398069 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4763ad3a-c76f-4750-8318-d18b1831a3e8-operator-scripts\") pod \"barbican-49a4-account-create-update-vmg9v\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.398084 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-config-data\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.398127 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl2dw\" (UniqueName: \"kubernetes.io/projected/4763ad3a-c76f-4750-8318-d18b1831a3e8-kube-api-access-nl2dw\") pod \"barbican-49a4-account-create-update-vmg9v\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.398154 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-combined-ca-bundle\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.399253 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7599e834-c07e-4db9-828b-675b53df2aa6-operator-scripts\") pod \"barbican-db-create-pbdln\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.401210 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-djdx9"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.402509 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.411790 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-djdx9"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.422514 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66v94\" (UniqueName: \"kubernetes.io/projected/7599e834-c07e-4db9-828b-675b53df2aa6-kube-api-access-66v94\") pod \"barbican-db-create-pbdln\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.442973 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.496565 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-392c-account-create-update-bv6kv"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.499321 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.500829 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t8wd\" (UniqueName: \"kubernetes.io/projected/493a9cc0-bb5a-463c-be84-c6def176bac7-kube-api-access-2t8wd\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501027 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4763ad3a-c76f-4750-8318-d18b1831a3e8-operator-scripts\") pod \"barbican-49a4-account-create-update-vmg9v\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501110 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-config-data\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501199 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5j4z\" (UniqueName: \"kubernetes.io/projected/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-kube-api-access-v5j4z\") pod \"neutron-db-create-djdx9\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501388 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3c64713-aa61-41bf-9e97-4edc35292d40-operator-scripts\") pod \"heat-db-create-tncs8\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501490 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-operator-scripts\") pod \"neutron-db-create-djdx9\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501583 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl2dw\" (UniqueName: \"kubernetes.io/projected/4763ad3a-c76f-4750-8318-d18b1831a3e8-kube-api-access-nl2dw\") pod \"barbican-49a4-account-create-update-vmg9v\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501686 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-combined-ca-bundle\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.501869 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddksc\" (UniqueName: \"kubernetes.io/projected/d3c64713-aa61-41bf-9e97-4edc35292d40-kube-api-access-ddksc\") pod \"heat-db-create-tncs8\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.503519 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4763ad3a-c76f-4750-8318-d18b1831a3e8-operator-scripts\") pod \"barbican-49a4-account-create-update-vmg9v\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.504044 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.506481 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-config-data\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.511481 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-392c-account-create-update-bv6kv"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.524637 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-combined-ca-bundle\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.525638 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl2dw\" (UniqueName: \"kubernetes.io/projected/4763ad3a-c76f-4750-8318-d18b1831a3e8-kube-api-access-nl2dw\") pod \"barbican-49a4-account-create-update-vmg9v\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.554340 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t8wd\" (UniqueName: \"kubernetes.io/projected/493a9cc0-bb5a-463c-be84-c6def176bac7-kube-api-access-2t8wd\") pod \"keystone-db-sync-gzplt\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.587382 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.589644 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b6c2-account-create-update-f4v77"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.590708 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.598285 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603012 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-operator-scripts\") pod \"neutron-db-create-djdx9\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603099 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddksc\" (UniqueName: \"kubernetes.io/projected/d3c64713-aa61-41bf-9e97-4edc35292d40-kube-api-access-ddksc\") pod \"heat-db-create-tncs8\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603173 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5j4z\" (UniqueName: \"kubernetes.io/projected/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-kube-api-access-v5j4z\") pod \"neutron-db-create-djdx9\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603201 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fd26d47-662e-4943-ac9d-fae6e8f75a81-operator-scripts\") pod \"heat-392c-account-create-update-bv6kv\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603220 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t4lm\" (UniqueName: \"kubernetes.io/projected/5fd26d47-662e-4943-ac9d-fae6e8f75a81-kube-api-access-5t4lm\") pod \"heat-392c-account-create-update-bv6kv\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603240 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3c64713-aa61-41bf-9e97-4edc35292d40-operator-scripts\") pod \"heat-db-create-tncs8\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.603941 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3c64713-aa61-41bf-9e97-4edc35292d40-operator-scripts\") pod \"heat-db-create-tncs8\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.604464 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-operator-scripts\") pod \"neutron-db-create-djdx9\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.621637 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b6c2-account-create-update-f4v77"] Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.625412 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddksc\" (UniqueName: \"kubernetes.io/projected/d3c64713-aa61-41bf-9e97-4edc35292d40-kube-api-access-ddksc\") pod \"heat-db-create-tncs8\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.634693 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5j4z\" (UniqueName: \"kubernetes.io/projected/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-kube-api-access-v5j4z\") pod \"neutron-db-create-djdx9\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.704257 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.704732 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d46f3006-1347-4670-ab14-1901562df15a-operator-scripts\") pod \"neutron-b6c2-account-create-update-f4v77\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.704821 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5mnc\" (UniqueName: \"kubernetes.io/projected/d46f3006-1347-4670-ab14-1901562df15a-kube-api-access-b5mnc\") pod \"neutron-b6c2-account-create-update-f4v77\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.704924 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fd26d47-662e-4943-ac9d-fae6e8f75a81-operator-scripts\") pod \"heat-392c-account-create-update-bv6kv\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.704955 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t4lm\" (UniqueName: \"kubernetes.io/projected/5fd26d47-662e-4943-ac9d-fae6e8f75a81-kube-api-access-5t4lm\") pod \"heat-392c-account-create-update-bv6kv\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.705826 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fd26d47-662e-4943-ac9d-fae6e8f75a81-operator-scripts\") pod \"heat-392c-account-create-update-bv6kv\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.722288 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t4lm\" (UniqueName: \"kubernetes.io/projected/5fd26d47-662e-4943-ac9d-fae6e8f75a81-kube-api-access-5t4lm\") pod \"heat-392c-account-create-update-bv6kv\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.722646 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tncs8" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.767261 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.806057 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d46f3006-1347-4670-ab14-1901562df15a-operator-scripts\") pod \"neutron-b6c2-account-create-update-f4v77\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.806138 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5mnc\" (UniqueName: \"kubernetes.io/projected/d46f3006-1347-4670-ab14-1901562df15a-kube-api-access-b5mnc\") pod \"neutron-b6c2-account-create-update-f4v77\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.807115 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d46f3006-1347-4670-ab14-1901562df15a-operator-scripts\") pod \"neutron-b6c2-account-create-update-f4v77\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.835496 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5mnc\" (UniqueName: \"kubernetes.io/projected/d46f3006-1347-4670-ab14-1901562df15a-kube-api-access-b5mnc\") pod \"neutron-b6c2-account-create-update-f4v77\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.891282 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:07 crc kubenswrapper[4817]: I1128 14:46:07.919172 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.440251 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571481 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqjhz\" (UniqueName: \"kubernetes.io/projected/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-kube-api-access-cqjhz\") pod \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571520 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run-ovn\") pod \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571536 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-log-ovn\") pod \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571585 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-additional-scripts\") pod \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571631 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-scripts\") pod \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571761 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run\") pod \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\" (UID: \"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13\") " Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.571910 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" (UID: "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.572143 4817 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.572167 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run" (OuterVolumeSpecName: "var-run") pod "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" (UID: "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.572623 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" (UID: "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.572817 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" (UID: "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.572957 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-scripts" (OuterVolumeSpecName: "scripts") pod "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" (UID: "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.614979 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-kube-api-access-cqjhz" (OuterVolumeSpecName: "kube-api-access-cqjhz") pod "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" (UID: "92e9f311-26cd-4c4b-9aa6-d3255ccb2d13"). InnerVolumeSpecName "kube-api-access-cqjhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.625811 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pwgz2-config-shlhj" event={"ID":"92e9f311-26cd-4c4b-9aa6-d3255ccb2d13","Type":"ContainerDied","Data":"d84727f28322476525ecba480dc0e76d0bdff462019470251aa9cb70976ceed6"} Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.625845 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d84727f28322476525ecba480dc0e76d0bdff462019470251aa9cb70976ceed6" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.625901 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pwgz2-config-shlhj" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.675736 4817 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.675763 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqjhz\" (UniqueName: \"kubernetes.io/projected/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-kube-api-access-cqjhz\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.675773 4817 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.675784 4817 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:11 crc kubenswrapper[4817]: I1128 14:46:11.675795 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.393057 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f9d9-account-create-update-v74l9"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.547398 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pwgz2-config-shlhj"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.555559 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pwgz2-config-shlhj"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.636813 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f9d9-account-create-update-v74l9" event={"ID":"eb9853b7-9f8c-4b0c-8b02-4bf014636556","Type":"ContainerStarted","Data":"bd561cb9670a35b20e8bf3308880d8baaad2ca4a433902878514752ddf763124"} Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.766411 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-49a4-account-create-update-vmg9v"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.776537 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-392c-account-create-update-bv6kv"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.782785 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gzplt"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.809111 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-tncs8"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.818331 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b6c2-account-create-update-f4v77"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.823931 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5p8hr"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.842871 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pbdln"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.849665 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-djdx9"] Nov 28 14:46:12 crc kubenswrapper[4817]: I1128 14:46:12.961119 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.658894 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-djdx9" event={"ID":"6d1e6a5f-677d-43bc-9790-9f85ade5db4d","Type":"ContainerStarted","Data":"fe79c77f6be1f5f7abf91e827e594e5965352e177544ef92118ce8d281c0bf06"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.659802 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5p8hr" event={"ID":"56eb98d9-0aff-4113-b6f0-34329d3b476d","Type":"ContainerStarted","Data":"3a193f705b15eef09fce0edb9c55e268afbfa2790b56852cc0eb2dcba0c9e0a5"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.660530 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pbdln" event={"ID":"7599e834-c07e-4db9-828b-675b53df2aa6","Type":"ContainerStarted","Data":"0d479ff3e79a7e86823708bd1411710be8bd8546c9269ffce79bd83ccb73f05e"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.661431 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-49a4-account-create-update-vmg9v" event={"ID":"4763ad3a-c76f-4750-8318-d18b1831a3e8","Type":"ContainerStarted","Data":"32c9d5829dcfe342f0300b8ed01d61cbd896d6ebdbca3fb094a7108497afdee0"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.662672 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gzplt" event={"ID":"493a9cc0-bb5a-463c-be84-c6def176bac7","Type":"ContainerStarted","Data":"cb178b92ea20d6a4413afb604cb88d3d249a441117eb0d80e5ecddba6e55fea8"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.665797 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-392c-account-create-update-bv6kv" event={"ID":"5fd26d47-662e-4943-ac9d-fae6e8f75a81","Type":"ContainerStarted","Data":"810b79c58940272462dfea9f2ab535926789bb7c48a17cbd7ce1c47cb40377ad"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.667375 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"c74a0989e4d491250e342770cb8ddea7855823e62bec335882a4ea8c270c393b"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.668534 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b6c2-account-create-update-f4v77" event={"ID":"d46f3006-1347-4670-ab14-1901562df15a","Type":"ContainerStarted","Data":"430b3d5496f1ae70721c739ef0417a4a6f332fcca335d30d404cfb8900771a5a"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.670247 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-tncs8" event={"ID":"d3c64713-aa61-41bf-9e97-4edc35292d40","Type":"ContainerStarted","Data":"9138909bc6162cfa24439b0dfb4da48cb2d37a4f24962d00de7b7d1f1fe325b0"} Nov 28 14:46:13 crc kubenswrapper[4817]: I1128 14:46:13.759548 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" path="/var/lib/kubelet/pods/92e9f311-26cd-4c4b-9aa6-d3255ccb2d13/volumes" Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.679869 4817 generic.go:334] "Generic (PLEG): container finished" podID="5fd26d47-662e-4943-ac9d-fae6e8f75a81" containerID="8ef295d27a65648074178bfa9277675404a7fa65b59f04a5fbb37733f2790a0d" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.679983 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-392c-account-create-update-bv6kv" event={"ID":"5fd26d47-662e-4943-ac9d-fae6e8f75a81","Type":"ContainerDied","Data":"8ef295d27a65648074178bfa9277675404a7fa65b59f04a5fbb37733f2790a0d"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.681694 4817 generic.go:334] "Generic (PLEG): container finished" podID="56eb98d9-0aff-4113-b6f0-34329d3b476d" containerID="b7dfc22974a5c6af3004cfcd83ea18cce90abbb2e1517132669651b7eba7f60b" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.681752 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5p8hr" event={"ID":"56eb98d9-0aff-4113-b6f0-34329d3b476d","Type":"ContainerDied","Data":"b7dfc22974a5c6af3004cfcd83ea18cce90abbb2e1517132669651b7eba7f60b"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.683043 4817 generic.go:334] "Generic (PLEG): container finished" podID="7599e834-c07e-4db9-828b-675b53df2aa6" containerID="6794b42950e35f9c2253282e6bf0a236d3a1ae64841d88538063f849c4f020fc" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.683087 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pbdln" event={"ID":"7599e834-c07e-4db9-828b-675b53df2aa6","Type":"ContainerDied","Data":"6794b42950e35f9c2253282e6bf0a236d3a1ae64841d88538063f849c4f020fc"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.689335 4817 generic.go:334] "Generic (PLEG): container finished" podID="4763ad3a-c76f-4750-8318-d18b1831a3e8" containerID="6deb01ae50ab392790d63376f8ce2bea35d7eb657a4953fedcc7edd7af4eb0f6" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.689382 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-49a4-account-create-update-vmg9v" event={"ID":"4763ad3a-c76f-4750-8318-d18b1831a3e8","Type":"ContainerDied","Data":"6deb01ae50ab392790d63376f8ce2bea35d7eb657a4953fedcc7edd7af4eb0f6"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.691903 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5gzqs" event={"ID":"77135fc7-4af7-4119-9153-525f2e7862c4","Type":"ContainerStarted","Data":"37d82d58bc3090b9566d85a315d65786e873485708b662958686d1bdb46af9d9"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.694144 4817 generic.go:334] "Generic (PLEG): container finished" podID="6d1e6a5f-677d-43bc-9790-9f85ade5db4d" containerID="4c9c61a9d17a53afd59ecb4c1ba414ab39925af267cb37f8992436ba5b1b16a6" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.694191 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-djdx9" event={"ID":"6d1e6a5f-677d-43bc-9790-9f85ade5db4d","Type":"ContainerDied","Data":"4c9c61a9d17a53afd59ecb4c1ba414ab39925af267cb37f8992436ba5b1b16a6"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.696473 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb9853b7-9f8c-4b0c-8b02-4bf014636556" containerID="d0d85161488f3692438f96efeda13be57155cde41e7bb429925c0ff295ef0da7" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.696626 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f9d9-account-create-update-v74l9" event={"ID":"eb9853b7-9f8c-4b0c-8b02-4bf014636556","Type":"ContainerDied","Data":"d0d85161488f3692438f96efeda13be57155cde41e7bb429925c0ff295ef0da7"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.699323 4817 generic.go:334] "Generic (PLEG): container finished" podID="d46f3006-1347-4670-ab14-1901562df15a" containerID="aff87f904424add4328aa2d941ae88c705fa26e457ebbf6614ebb965b4663553" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.699379 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b6c2-account-create-update-f4v77" event={"ID":"d46f3006-1347-4670-ab14-1901562df15a","Type":"ContainerDied","Data":"aff87f904424add4328aa2d941ae88c705fa26e457ebbf6614ebb965b4663553"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.701322 4817 generic.go:334] "Generic (PLEG): container finished" podID="d3c64713-aa61-41bf-9e97-4edc35292d40" containerID="1b312db9d53b12e3fe3eadd74e468e16d43c3db1cf8d0d61229e80863e2a8ac1" exitCode=0 Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.701389 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-tncs8" event={"ID":"d3c64713-aa61-41bf-9e97-4edc35292d40","Type":"ContainerDied","Data":"1b312db9d53b12e3fe3eadd74e468e16d43c3db1cf8d0d61229e80863e2a8ac1"} Nov 28 14:46:14 crc kubenswrapper[4817]: I1128 14:46:14.774811 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-5gzqs" podStartSLOduration=4.581358095 podStartE2EDuration="19.774790676s" podCreationTimestamp="2025-11-28 14:45:55 +0000 UTC" firstStartedPulling="2025-11-28 14:45:56.428790797 +0000 UTC m=+979.016769063" lastFinishedPulling="2025-11-28 14:46:11.622223378 +0000 UTC m=+994.210201644" observedRunningTime="2025-11-28 14:46:14.764428818 +0000 UTC m=+997.352407104" watchObservedRunningTime="2025-11-28 14:46:14.774790676 +0000 UTC m=+997.362768972" Nov 28 14:46:15 crc kubenswrapper[4817]: I1128 14:46:15.714320 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"6f05bed470109ce740c253438485b99a4ff7eab623398df09d38c5be7a963cf8"} Nov 28 14:46:15 crc kubenswrapper[4817]: I1128 14:46:15.714372 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"ea3d8644843382ac5dbe89a231917c241a461d0e80ed98b2c250ff2a343f7f99"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.254559 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.320777 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.329135 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.370226 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.394658 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.400921 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.409783 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.425960 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tncs8" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.435867 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcq5k\" (UniqueName: \"kubernetes.io/projected/56eb98d9-0aff-4113-b6f0-34329d3b476d-kube-api-access-lcq5k\") pod \"56eb98d9-0aff-4113-b6f0-34329d3b476d\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436013 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fd26d47-662e-4943-ac9d-fae6e8f75a81-operator-scripts\") pod \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436055 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t4lm\" (UniqueName: \"kubernetes.io/projected/5fd26d47-662e-4943-ac9d-fae6e8f75a81-kube-api-access-5t4lm\") pod \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\" (UID: \"5fd26d47-662e-4943-ac9d-fae6e8f75a81\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436108 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-operator-scripts\") pod \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436375 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56eb98d9-0aff-4113-b6f0-34329d3b476d-operator-scripts\") pod \"56eb98d9-0aff-4113-b6f0-34329d3b476d\" (UID: \"56eb98d9-0aff-4113-b6f0-34329d3b476d\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436419 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4763ad3a-c76f-4750-8318-d18b1831a3e8-operator-scripts\") pod \"4763ad3a-c76f-4750-8318-d18b1831a3e8\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436443 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5j4z\" (UniqueName: \"kubernetes.io/projected/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-kube-api-access-v5j4z\") pod \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\" (UID: \"6d1e6a5f-677d-43bc-9790-9f85ade5db4d\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.436462 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66v94\" (UniqueName: \"kubernetes.io/projected/7599e834-c07e-4db9-828b-675b53df2aa6-kube-api-access-66v94\") pod \"7599e834-c07e-4db9-828b-675b53df2aa6\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.439587 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56eb98d9-0aff-4113-b6f0-34329d3b476d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56eb98d9-0aff-4113-b6f0-34329d3b476d" (UID: "56eb98d9-0aff-4113-b6f0-34329d3b476d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.440914 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fd26d47-662e-4943-ac9d-fae6e8f75a81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5fd26d47-662e-4943-ac9d-fae6e8f75a81" (UID: "5fd26d47-662e-4943-ac9d-fae6e8f75a81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.443193 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d1e6a5f-677d-43bc-9790-9f85ade5db4d" (UID: "6d1e6a5f-677d-43bc-9790-9f85ade5db4d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.444168 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7599e834-c07e-4db9-828b-675b53df2aa6-kube-api-access-66v94" (OuterVolumeSpecName: "kube-api-access-66v94") pod "7599e834-c07e-4db9-828b-675b53df2aa6" (UID: "7599e834-c07e-4db9-828b-675b53df2aa6"). InnerVolumeSpecName "kube-api-access-66v94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.444223 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4763ad3a-c76f-4750-8318-d18b1831a3e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4763ad3a-c76f-4750-8318-d18b1831a3e8" (UID: "4763ad3a-c76f-4750-8318-d18b1831a3e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.445052 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd26d47-662e-4943-ac9d-fae6e8f75a81-kube-api-access-5t4lm" (OuterVolumeSpecName: "kube-api-access-5t4lm") pod "5fd26d47-662e-4943-ac9d-fae6e8f75a81" (UID: "5fd26d47-662e-4943-ac9d-fae6e8f75a81"). InnerVolumeSpecName "kube-api-access-5t4lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.446041 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-kube-api-access-v5j4z" (OuterVolumeSpecName: "kube-api-access-v5j4z") pod "6d1e6a5f-677d-43bc-9790-9f85ade5db4d" (UID: "6d1e6a5f-677d-43bc-9790-9f85ade5db4d"). InnerVolumeSpecName "kube-api-access-v5j4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.466420 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56eb98d9-0aff-4113-b6f0-34329d3b476d-kube-api-access-lcq5k" (OuterVolumeSpecName: "kube-api-access-lcq5k") pod "56eb98d9-0aff-4113-b6f0-34329d3b476d" (UID: "56eb98d9-0aff-4113-b6f0-34329d3b476d"). InnerVolumeSpecName "kube-api-access-lcq5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.537953 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl2dw\" (UniqueName: \"kubernetes.io/projected/4763ad3a-c76f-4750-8318-d18b1831a3e8-kube-api-access-nl2dw\") pod \"4763ad3a-c76f-4750-8318-d18b1831a3e8\" (UID: \"4763ad3a-c76f-4750-8318-d18b1831a3e8\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538014 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5mnc\" (UniqueName: \"kubernetes.io/projected/d46f3006-1347-4670-ab14-1901562df15a-kube-api-access-b5mnc\") pod \"d46f3006-1347-4670-ab14-1901562df15a\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538127 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddksc\" (UniqueName: \"kubernetes.io/projected/d3c64713-aa61-41bf-9e97-4edc35292d40-kube-api-access-ddksc\") pod \"d3c64713-aa61-41bf-9e97-4edc35292d40\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538171 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7599e834-c07e-4db9-828b-675b53df2aa6-operator-scripts\") pod \"7599e834-c07e-4db9-828b-675b53df2aa6\" (UID: \"7599e834-c07e-4db9-828b-675b53df2aa6\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538197 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3c64713-aa61-41bf-9e97-4edc35292d40-operator-scripts\") pod \"d3c64713-aa61-41bf-9e97-4edc35292d40\" (UID: \"d3c64713-aa61-41bf-9e97-4edc35292d40\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538221 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d46f3006-1347-4670-ab14-1901562df15a-operator-scripts\") pod \"d46f3006-1347-4670-ab14-1901562df15a\" (UID: \"d46f3006-1347-4670-ab14-1901562df15a\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538254 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb9853b7-9f8c-4b0c-8b02-4bf014636556-operator-scripts\") pod \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538320 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdrbn\" (UniqueName: \"kubernetes.io/projected/eb9853b7-9f8c-4b0c-8b02-4bf014636556-kube-api-access-mdrbn\") pod \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\" (UID: \"eb9853b7-9f8c-4b0c-8b02-4bf014636556\") " Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538604 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5fd26d47-662e-4943-ac9d-fae6e8f75a81-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538618 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t4lm\" (UniqueName: \"kubernetes.io/projected/5fd26d47-662e-4943-ac9d-fae6e8f75a81-kube-api-access-5t4lm\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538630 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538641 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56eb98d9-0aff-4113-b6f0-34329d3b476d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538652 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4763ad3a-c76f-4750-8318-d18b1831a3e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538664 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5j4z\" (UniqueName: \"kubernetes.io/projected/6d1e6a5f-677d-43bc-9790-9f85ade5db4d-kube-api-access-v5j4z\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538675 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66v94\" (UniqueName: \"kubernetes.io/projected/7599e834-c07e-4db9-828b-675b53df2aa6-kube-api-access-66v94\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538685 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcq5k\" (UniqueName: \"kubernetes.io/projected/56eb98d9-0aff-4113-b6f0-34329d3b476d-kube-api-access-lcq5k\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.538810 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7599e834-c07e-4db9-828b-675b53df2aa6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7599e834-c07e-4db9-828b-675b53df2aa6" (UID: "7599e834-c07e-4db9-828b-675b53df2aa6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.540188 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d46f3006-1347-4670-ab14-1901562df15a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d46f3006-1347-4670-ab14-1901562df15a" (UID: "d46f3006-1347-4670-ab14-1901562df15a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.540212 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3c64713-aa61-41bf-9e97-4edc35292d40-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3c64713-aa61-41bf-9e97-4edc35292d40" (UID: "d3c64713-aa61-41bf-9e97-4edc35292d40"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.542082 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb9853b7-9f8c-4b0c-8b02-4bf014636556-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb9853b7-9f8c-4b0c-8b02-4bf014636556" (UID: "eb9853b7-9f8c-4b0c-8b02-4bf014636556"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.542552 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9853b7-9f8c-4b0c-8b02-4bf014636556-kube-api-access-mdrbn" (OuterVolumeSpecName: "kube-api-access-mdrbn") pod "eb9853b7-9f8c-4b0c-8b02-4bf014636556" (UID: "eb9853b7-9f8c-4b0c-8b02-4bf014636556"). InnerVolumeSpecName "kube-api-access-mdrbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.543809 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46f3006-1347-4670-ab14-1901562df15a-kube-api-access-b5mnc" (OuterVolumeSpecName: "kube-api-access-b5mnc") pod "d46f3006-1347-4670-ab14-1901562df15a" (UID: "d46f3006-1347-4670-ab14-1901562df15a"). InnerVolumeSpecName "kube-api-access-b5mnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.543874 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4763ad3a-c76f-4750-8318-d18b1831a3e8-kube-api-access-nl2dw" (OuterVolumeSpecName: "kube-api-access-nl2dw") pod "4763ad3a-c76f-4750-8318-d18b1831a3e8" (UID: "4763ad3a-c76f-4750-8318-d18b1831a3e8"). InnerVolumeSpecName "kube-api-access-nl2dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.544857 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3c64713-aa61-41bf-9e97-4edc35292d40-kube-api-access-ddksc" (OuterVolumeSpecName: "kube-api-access-ddksc") pod "d3c64713-aa61-41bf-9e97-4edc35292d40" (UID: "d3c64713-aa61-41bf-9e97-4edc35292d40"). InnerVolumeSpecName "kube-api-access-ddksc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.645981 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5mnc\" (UniqueName: \"kubernetes.io/projected/d46f3006-1347-4670-ab14-1901562df15a-kube-api-access-b5mnc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646415 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddksc\" (UniqueName: \"kubernetes.io/projected/d3c64713-aa61-41bf-9e97-4edc35292d40-kube-api-access-ddksc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646428 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7599e834-c07e-4db9-828b-675b53df2aa6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646440 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3c64713-aa61-41bf-9e97-4edc35292d40-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646453 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d46f3006-1347-4670-ab14-1901562df15a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646464 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb9853b7-9f8c-4b0c-8b02-4bf014636556-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646475 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdrbn\" (UniqueName: \"kubernetes.io/projected/eb9853b7-9f8c-4b0c-8b02-4bf014636556-kube-api-access-mdrbn\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.646488 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl2dw\" (UniqueName: \"kubernetes.io/projected/4763ad3a-c76f-4750-8318-d18b1831a3e8-kube-api-access-nl2dw\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.763199 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-392c-account-create-update-bv6kv" event={"ID":"5fd26d47-662e-4943-ac9d-fae6e8f75a81","Type":"ContainerDied","Data":"810b79c58940272462dfea9f2ab535926789bb7c48a17cbd7ce1c47cb40377ad"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.763447 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="810b79c58940272462dfea9f2ab535926789bb7c48a17cbd7ce1c47cb40377ad" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.763370 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-392c-account-create-update-bv6kv" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.771282 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"cd8f36136dd2dc92c211f0c127bc0aafa8111450c1833b762455222e61e538c8"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.773419 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gzplt" event={"ID":"493a9cc0-bb5a-463c-be84-c6def176bac7","Type":"ContainerStarted","Data":"1694bf4ac9a28943f0d84e487884ffe418ebf912fbe4706753797183e070c064"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.775309 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f9d9-account-create-update-v74l9" event={"ID":"eb9853b7-9f8c-4b0c-8b02-4bf014636556","Type":"ContainerDied","Data":"bd561cb9670a35b20e8bf3308880d8baaad2ca4a433902878514752ddf763124"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.775353 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd561cb9670a35b20e8bf3308880d8baaad2ca4a433902878514752ddf763124" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.775424 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f9d9-account-create-update-v74l9" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.780838 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5p8hr" event={"ID":"56eb98d9-0aff-4113-b6f0-34329d3b476d","Type":"ContainerDied","Data":"3a193f705b15eef09fce0edb9c55e268afbfa2790b56852cc0eb2dcba0c9e0a5"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.780886 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a193f705b15eef09fce0edb9c55e268afbfa2790b56852cc0eb2dcba0c9e0a5" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.781342 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5p8hr" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.785731 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pbdln" event={"ID":"7599e834-c07e-4db9-828b-675b53df2aa6","Type":"ContainerDied","Data":"0d479ff3e79a7e86823708bd1411710be8bd8546c9269ffce79bd83ccb73f05e"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.785754 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d479ff3e79a7e86823708bd1411710be8bd8546c9269ffce79bd83ccb73f05e" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.785811 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pbdln" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.792579 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-tncs8" event={"ID":"d3c64713-aa61-41bf-9e97-4edc35292d40","Type":"ContainerDied","Data":"9138909bc6162cfa24439b0dfb4da48cb2d37a4f24962d00de7b7d1f1fe325b0"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.792646 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9138909bc6162cfa24439b0dfb4da48cb2d37a4f24962d00de7b7d1f1fe325b0" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.792599 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tncs8" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.796339 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-49a4-account-create-update-vmg9v" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.796351 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-49a4-account-create-update-vmg9v" event={"ID":"4763ad3a-c76f-4750-8318-d18b1831a3e8","Type":"ContainerDied","Data":"32c9d5829dcfe342f0300b8ed01d61cbd896d6ebdbca3fb094a7108497afdee0"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.796386 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32c9d5829dcfe342f0300b8ed01d61cbd896d6ebdbca3fb094a7108497afdee0" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.799643 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-djdx9" event={"ID":"6d1e6a5f-677d-43bc-9790-9f85ade5db4d","Type":"ContainerDied","Data":"fe79c77f6be1f5f7abf91e827e594e5965352e177544ef92118ce8d281c0bf06"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.799671 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe79c77f6be1f5f7abf91e827e594e5965352e177544ef92118ce8d281c0bf06" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.799745 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-djdx9" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.806780 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b6c2-account-create-update-f4v77" event={"ID":"d46f3006-1347-4670-ab14-1901562df15a","Type":"ContainerDied","Data":"430b3d5496f1ae70721c739ef0417a4a6f332fcca335d30d404cfb8900771a5a"} Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.806818 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="430b3d5496f1ae70721c739ef0417a4a6f332fcca335d30d404cfb8900771a5a" Nov 28 14:46:19 crc kubenswrapper[4817]: I1128 14:46:19.806824 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b6c2-account-create-update-f4v77" Nov 28 14:46:20 crc kubenswrapper[4817]: I1128 14:46:20.302240 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gzplt" podStartSLOduration=7.0604850710000004 podStartE2EDuration="13.302175675s" podCreationTimestamp="2025-11-28 14:46:07 +0000 UTC" firstStartedPulling="2025-11-28 14:46:12.804665726 +0000 UTC m=+995.392644002" lastFinishedPulling="2025-11-28 14:46:19.04635634 +0000 UTC m=+1001.634334606" observedRunningTime="2025-11-28 14:46:19.792747094 +0000 UTC m=+1002.380725380" watchObservedRunningTime="2025-11-28 14:46:20.302175675 +0000 UTC m=+1002.890153941" Nov 28 14:46:20 crc kubenswrapper[4817]: I1128 14:46:20.826708 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"78bad31337c7faaad7871abf81b111a86df63946e5f477a5f3f2ccdba6419ca3"} Nov 28 14:46:21 crc kubenswrapper[4817]: I1128 14:46:21.846967 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"2f301e390368fa3c3a41ec9f0665bc8c04edb5f1bdfffcc40d3b52e3c8c6a72c"} Nov 28 14:46:22 crc kubenswrapper[4817]: I1128 14:46:22.864405 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"5bf8cd6c1dfaa617a6fb0adb589e9b6d4e56985c5240c55892ca18e71232ee03"} Nov 28 14:46:22 crc kubenswrapper[4817]: I1128 14:46:22.864983 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"30d0aa4ee15fc395b23202043245a00e3f812867518d3fdc2c0e41ea978d2cd8"} Nov 28 14:46:22 crc kubenswrapper[4817]: I1128 14:46:22.864999 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"9efa227ed4f8d09a1cd0700bb82c5009050aba814c21fb5f0f7519cecabc7ba2"} Nov 28 14:46:23 crc kubenswrapper[4817]: I1128 14:46:23.874681 4817 generic.go:334] "Generic (PLEG): container finished" podID="77135fc7-4af7-4119-9153-525f2e7862c4" containerID="37d82d58bc3090b9566d85a315d65786e873485708b662958686d1bdb46af9d9" exitCode=0 Nov 28 14:46:23 crc kubenswrapper[4817]: I1128 14:46:23.874782 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5gzqs" event={"ID":"77135fc7-4af7-4119-9153-525f2e7862c4","Type":"ContainerDied","Data":"37d82d58bc3090b9566d85a315d65786e873485708b662958686d1bdb46af9d9"} Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.047177 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.047231 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.883135 4817 generic.go:334] "Generic (PLEG): container finished" podID="493a9cc0-bb5a-463c-be84-c6def176bac7" containerID="1694bf4ac9a28943f0d84e487884ffe418ebf912fbe4706753797183e070c064" exitCode=0 Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.883225 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gzplt" event={"ID":"493a9cc0-bb5a-463c-be84-c6def176bac7","Type":"ContainerDied","Data":"1694bf4ac9a28943f0d84e487884ffe418ebf912fbe4706753797183e070c064"} Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.891654 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"551956f7bab9aa2d19f2ae22de6430e24f5c750806747ff95cc097c8b460667b"} Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.891691 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"16ecd496f2de10b7bee2d975fe27e368d9bd4c47cdad3db6f4f7f888b9fd8187"} Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.892516 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"768493fc09cc86bda49c2963a2363e8d1f43156d80fd438d48e0278d30f9148a"} Nov 28 14:46:24 crc kubenswrapper[4817]: I1128 14:46:24.892530 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"acac807482f13dcf8870dbde3ee73b7b9acfd40645da4aeab401881c601a9adc"} Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.351158 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5gzqs" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.467319 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66s8r\" (UniqueName: \"kubernetes.io/projected/77135fc7-4af7-4119-9153-525f2e7862c4-kube-api-access-66s8r\") pod \"77135fc7-4af7-4119-9153-525f2e7862c4\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.467389 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-config-data\") pod \"77135fc7-4af7-4119-9153-525f2e7862c4\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.467411 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-db-sync-config-data\") pod \"77135fc7-4af7-4119-9153-525f2e7862c4\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.467541 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-combined-ca-bundle\") pod \"77135fc7-4af7-4119-9153-525f2e7862c4\" (UID: \"77135fc7-4af7-4119-9153-525f2e7862c4\") " Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.474895 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "77135fc7-4af7-4119-9153-525f2e7862c4" (UID: "77135fc7-4af7-4119-9153-525f2e7862c4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.474985 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77135fc7-4af7-4119-9153-525f2e7862c4-kube-api-access-66s8r" (OuterVolumeSpecName: "kube-api-access-66s8r") pod "77135fc7-4af7-4119-9153-525f2e7862c4" (UID: "77135fc7-4af7-4119-9153-525f2e7862c4"). InnerVolumeSpecName "kube-api-access-66s8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.493765 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77135fc7-4af7-4119-9153-525f2e7862c4" (UID: "77135fc7-4af7-4119-9153-525f2e7862c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.531486 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-config-data" (OuterVolumeSpecName: "config-data") pod "77135fc7-4af7-4119-9153-525f2e7862c4" (UID: "77135fc7-4af7-4119-9153-525f2e7862c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.568948 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.568977 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66s8r\" (UniqueName: \"kubernetes.io/projected/77135fc7-4af7-4119-9153-525f2e7862c4-kube-api-access-66s8r\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.568991 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.569001 4817 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77135fc7-4af7-4119-9153-525f2e7862c4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.901184 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5gzqs" event={"ID":"77135fc7-4af7-4119-9153-525f2e7862c4","Type":"ContainerDied","Data":"0107cc5b36e5fad4599d78db2823d09c2275d7a97c6cfa1863565f325fe8d786"} Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.901239 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0107cc5b36e5fad4599d78db2823d09c2275d7a97c6cfa1863565f325fe8d786" Nov 28 14:46:25 crc kubenswrapper[4817]: I1128 14:46:25.901276 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5gzqs" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.310486 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-ps576"] Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311024 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1e6a5f-677d-43bc-9790-9f85ade5db4d" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311038 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1e6a5f-677d-43bc-9790-9f85ade5db4d" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311054 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd26d47-662e-4943-ac9d-fae6e8f75a81" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311061 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd26d47-662e-4943-ac9d-fae6e8f75a81" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311071 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77135fc7-4af7-4119-9153-525f2e7862c4" containerName="glance-db-sync" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311078 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="77135fc7-4af7-4119-9153-525f2e7862c4" containerName="glance-db-sync" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311084 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c64713-aa61-41bf-9e97-4edc35292d40" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311092 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c64713-aa61-41bf-9e97-4edc35292d40" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311109 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7599e834-c07e-4db9-828b-675b53df2aa6" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311114 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="7599e834-c07e-4db9-828b-675b53df2aa6" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311127 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4763ad3a-c76f-4750-8318-d18b1831a3e8" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311132 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="4763ad3a-c76f-4750-8318-d18b1831a3e8" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311139 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56eb98d9-0aff-4113-b6f0-34329d3b476d" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311144 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="56eb98d9-0aff-4113-b6f0-34329d3b476d" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311152 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" containerName="ovn-config" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311158 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" containerName="ovn-config" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311167 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d46f3006-1347-4670-ab14-1901562df15a" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311172 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d46f3006-1347-4670-ab14-1901562df15a" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: E1128 14:46:26.311183 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9853b7-9f8c-4b0c-8b02-4bf014636556" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311188 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9853b7-9f8c-4b0c-8b02-4bf014636556" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311335 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d46f3006-1347-4670-ab14-1901562df15a" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311358 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="7599e834-c07e-4db9-828b-675b53df2aa6" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311373 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3c64713-aa61-41bf-9e97-4edc35292d40" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311389 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e9f311-26cd-4c4b-9aa6-d3255ccb2d13" containerName="ovn-config" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311399 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9853b7-9f8c-4b0c-8b02-4bf014636556" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311409 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="77135fc7-4af7-4119-9153-525f2e7862c4" containerName="glance-db-sync" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311420 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="4763ad3a-c76f-4750-8318-d18b1831a3e8" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311429 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd26d47-662e-4943-ac9d-fae6e8f75a81" containerName="mariadb-account-create-update" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311438 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="56eb98d9-0aff-4113-b6f0-34329d3b476d" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.311445 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1e6a5f-677d-43bc-9790-9f85ade5db4d" containerName="mariadb-database-create" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.312246 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.326713 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-ps576"] Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.484497 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.484853 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb2tx\" (UniqueName: \"kubernetes.io/projected/fc32c327-72f5-444d-886e-b85a04ae31ba-kube-api-access-bb2tx\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.484953 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.485469 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.485595 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-config\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.587441 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-config\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.587590 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.587646 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb2tx\" (UniqueName: \"kubernetes.io/projected/fc32c327-72f5-444d-886e-b85a04ae31ba-kube-api-access-bb2tx\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.587691 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.587894 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.588651 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-config\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.588786 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.588921 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.589596 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.604919 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb2tx\" (UniqueName: \"kubernetes.io/projected/fc32c327-72f5-444d-886e-b85a04ae31ba-kube-api-access-bb2tx\") pod \"dnsmasq-dns-5b946c75cc-ps576\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:26 crc kubenswrapper[4817]: I1128 14:46:26.627607 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.097923 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-ps576"] Nov 28 14:46:27 crc kubenswrapper[4817]: W1128 14:46:27.102981 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc32c327_72f5_444d_886e_b85a04ae31ba.slice/crio-554c471a9730bf399e88a01f271691d5ed1ddb65c2ad604efda288a1dc037560 WatchSource:0}: Error finding container 554c471a9730bf399e88a01f271691d5ed1ddb65c2ad604efda288a1dc037560: Status 404 returned error can't find the container with id 554c471a9730bf399e88a01f271691d5ed1ddb65c2ad604efda288a1dc037560 Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.223378 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.301544 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-combined-ca-bundle\") pod \"493a9cc0-bb5a-463c-be84-c6def176bac7\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.301607 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-config-data\") pod \"493a9cc0-bb5a-463c-be84-c6def176bac7\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.301644 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t8wd\" (UniqueName: \"kubernetes.io/projected/493a9cc0-bb5a-463c-be84-c6def176bac7-kube-api-access-2t8wd\") pod \"493a9cc0-bb5a-463c-be84-c6def176bac7\" (UID: \"493a9cc0-bb5a-463c-be84-c6def176bac7\") " Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.309322 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493a9cc0-bb5a-463c-be84-c6def176bac7-kube-api-access-2t8wd" (OuterVolumeSpecName: "kube-api-access-2t8wd") pod "493a9cc0-bb5a-463c-be84-c6def176bac7" (UID: "493a9cc0-bb5a-463c-be84-c6def176bac7"). InnerVolumeSpecName "kube-api-access-2t8wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.325068 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "493a9cc0-bb5a-463c-be84-c6def176bac7" (UID: "493a9cc0-bb5a-463c-be84-c6def176bac7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.352887 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-config-data" (OuterVolumeSpecName: "config-data") pod "493a9cc0-bb5a-463c-be84-c6def176bac7" (UID: "493a9cc0-bb5a-463c-be84-c6def176bac7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.403177 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t8wd\" (UniqueName: \"kubernetes.io/projected/493a9cc0-bb5a-463c-be84-c6def176bac7-kube-api-access-2t8wd\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.403216 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.403229 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493a9cc0-bb5a-463c-be84-c6def176bac7-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.939980 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gzplt" event={"ID":"493a9cc0-bb5a-463c-be84-c6def176bac7","Type":"ContainerDied","Data":"cb178b92ea20d6a4413afb604cb88d3d249a441117eb0d80e5ecddba6e55fea8"} Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.940034 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb178b92ea20d6a4413afb604cb88d3d249a441117eb0d80e5ecddba6e55fea8" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.940010 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gzplt" Nov 28 14:46:27 crc kubenswrapper[4817]: I1128 14:46:27.941694 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" event={"ID":"fc32c327-72f5-444d-886e-b85a04ae31ba","Type":"ContainerStarted","Data":"554c471a9730bf399e88a01f271691d5ed1ddb65c2ad604efda288a1dc037560"} Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.570226 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-ps576"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.583562 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9lhpb"] Nov 28 14:46:28 crc kubenswrapper[4817]: E1128 14:46:28.583944 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493a9cc0-bb5a-463c-be84-c6def176bac7" containerName="keystone-db-sync" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.583960 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="493a9cc0-bb5a-463c-be84-c6def176bac7" containerName="keystone-db-sync" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.584130 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="493a9cc0-bb5a-463c-be84-c6def176bac7" containerName="keystone-db-sync" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.588086 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.591656 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.591942 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.592054 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.600269 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-knk6m" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.600736 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.612902 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9lhpb"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.626779 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784f69c749-jbnlp"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.628461 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.664653 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-jbnlp"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.711895 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-fbfcr"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.713185 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.723536 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-credential-keys\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.723568 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-config-data\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.723629 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-scripts\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.723674 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-combined-ca-bundle\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.723701 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvb5j\" (UniqueName: \"kubernetes.io/projected/71e2f1db-1c92-4c41-9ef1-92fc41189f43-kube-api-access-jvb5j\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.723750 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-fernet-keys\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.725250 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-jmpmj" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.731999 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.805324 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-fbfcr"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.833923 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-combined-ca-bundle\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834000 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-config\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834029 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvb5j\" (UniqueName: \"kubernetes.io/projected/71e2f1db-1c92-4c41-9ef1-92fc41189f43-kube-api-access-jvb5j\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834070 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834104 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-config-data\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834127 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-fernet-keys\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834178 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-credential-keys\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834207 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-config-data\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834231 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-combined-ca-bundle\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834256 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4g7\" (UniqueName: \"kubernetes.io/projected/6da990a2-1382-4707-8311-5dd46e4f537c-kube-api-access-dw4g7\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834313 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmmmn\" (UniqueName: \"kubernetes.io/projected/d69ea08f-64fd-49ca-be54-7dfc35fd5556-kube-api-access-cmmmn\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834334 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-scripts\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834354 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.834394 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-dns-svc\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.854015 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-credential-keys\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.854695 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-config-data\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.855476 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-fernet-keys\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.856174 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-combined-ca-bundle\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.857088 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-scripts\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.889573 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-5q2qk"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.890707 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.907030 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.907708 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dc9bg" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.908699 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.912047 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-fflrd"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.913249 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.914856 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvb5j\" (UniqueName: \"kubernetes.io/projected/71e2f1db-1c92-4c41-9ef1-92fc41189f43-kube-api-access-jvb5j\") pod \"keystone-bootstrap-9lhpb\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.918046 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9hvz9" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.918248 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.926828 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936611 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4g7\" (UniqueName: \"kubernetes.io/projected/6da990a2-1382-4707-8311-5dd46e4f537c-kube-api-access-dw4g7\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936698 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmmmn\" (UniqueName: \"kubernetes.io/projected/d69ea08f-64fd-49ca-be54-7dfc35fd5556-kube-api-access-cmmmn\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936737 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936768 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-dns-svc\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936803 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-config\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936836 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936856 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-config-data\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.936897 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-combined-ca-bundle\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.947315 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.947771 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-dns-svc\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.947855 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-config\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.953301 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.962882 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5q2qk"] Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.970165 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-combined-ca-bundle\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.970702 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-config-data\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.972046 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4g7\" (UniqueName: \"kubernetes.io/projected/6da990a2-1382-4707-8311-5dd46e4f537c-kube-api-access-dw4g7\") pod \"heat-db-sync-fbfcr\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.983519 4817 generic.go:334] "Generic (PLEG): container finished" podID="fc32c327-72f5-444d-886e-b85a04ae31ba" containerID="e994d4d685f639d0fb47040768a2611c6c86b975dec8e8284382c36424c1189d" exitCode=0 Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.983652 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" event={"ID":"fc32c327-72f5-444d-886e-b85a04ae31ba","Type":"ContainerDied","Data":"e994d4d685f639d0fb47040768a2611c6c86b975dec8e8284382c36424c1189d"} Nov 28 14:46:28 crc kubenswrapper[4817]: I1128 14:46:28.989911 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fflrd"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.030764 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"9eeb5c42a67190f6a8620a11ecb7ffca7dc0c4f64175b2d4c092546971910df0"} Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.032312 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmmmn\" (UniqueName: \"kubernetes.io/projected/d69ea08f-64fd-49ca-be54-7dfc35fd5556-kube-api-access-cmmmn\") pod \"dnsmasq-dns-784f69c749-jbnlp\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038266 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-scripts\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038332 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-config-data\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038394 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-config\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038449 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29sxk\" (UniqueName: \"kubernetes.io/projected/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-kube-api-access-29sxk\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038483 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-combined-ca-bundle\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038505 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-combined-ca-bundle\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038536 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-etc-machine-id\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.038560 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfctv\" (UniqueName: \"kubernetes.io/projected/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-kube-api-access-pfctv\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.055817 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gvbcw"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.055816 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-db-sync-config-data\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.063496 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fbfcr" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.084608 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.087427 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.091935 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-kcxkk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.092175 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.138570 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-mpzzx"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.140799 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.143239 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-t9vqz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.143541 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.164673 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-scripts\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.164978 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-config-data\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165095 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-config\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165206 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29sxk\" (UniqueName: \"kubernetes.io/projected/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-kube-api-access-29sxk\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165387 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-combined-ca-bundle\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165473 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-combined-ca-bundle\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165562 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-etc-machine-id\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165632 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfctv\" (UniqueName: \"kubernetes.io/projected/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-kube-api-access-pfctv\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.165730 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-db-sync-config-data\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.169838 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-etc-machine-id\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.176485 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-db-sync-config-data\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.177188 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-scripts\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.177575 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-config\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.178258 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-config-data\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.178926 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-combined-ca-bundle\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.179975 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-combined-ca-bundle\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.197452 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29sxk\" (UniqueName: \"kubernetes.io/projected/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-kube-api-access-29sxk\") pod \"neutron-db-sync-fflrd\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.197993 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gvbcw"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.202369 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfctv\" (UniqueName: \"kubernetes.io/projected/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-kube-api-access-pfctv\") pod \"cinder-db-sync-5q2qk\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.212557 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.223435 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mpzzx"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.236776 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-jbnlp"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.237418 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.268462 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-db-sync-config-data\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.268835 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkg8l\" (UniqueName: \"kubernetes.io/projected/11e39ba1-49a5-4289-a913-10b87e82f4e0-kube-api-access-hkg8l\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.268857 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-config-data\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.268873 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-combined-ca-bundle\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.268953 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-scripts\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.268976 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhhqd\" (UniqueName: \"kubernetes.io/projected/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-kube-api-access-jhhqd\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.269008 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-combined-ca-bundle\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.269055 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-logs\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.288911 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-pwmxz"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.295244 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.308379 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.310314 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.315252 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.315461 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.325917 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-pwmxz"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.333470 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371740 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-logs\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371793 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-dns-svc\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371831 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-db-sync-config-data\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371848 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371889 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkg8l\" (UniqueName: \"kubernetes.io/projected/11e39ba1-49a5-4289-a913-10b87e82f4e0-kube-api-access-hkg8l\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371941 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-config-data\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.371963 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-combined-ca-bundle\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.372013 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-config\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.372038 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.372065 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrj8g\" (UniqueName: \"kubernetes.io/projected/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-kube-api-access-jrj8g\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.372085 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-scripts\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.372110 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhhqd\" (UniqueName: \"kubernetes.io/projected/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-kube-api-access-jhhqd\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.372139 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-combined-ca-bundle\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.378597 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-logs\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.379333 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-config-data\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.383003 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-combined-ca-bundle\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.393262 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-combined-ca-bundle\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.402872 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-scripts\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.402952 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-db-sync-config-data\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.415559 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkg8l\" (UniqueName: \"kubernetes.io/projected/11e39ba1-49a5-4289-a913-10b87e82f4e0-kube-api-access-hkg8l\") pod \"barbican-db-sync-mpzzx\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.415596 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhhqd\" (UniqueName: \"kubernetes.io/projected/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-kube-api-access-jhhqd\") pod \"placement-db-sync-gvbcw\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.415801 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fflrd" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.438219 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.469183 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvbcw" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473156 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx8qd\" (UniqueName: \"kubernetes.io/projected/49a012fe-0369-4c46-89e8-0c0062f46c33-kube-api-access-gx8qd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473201 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-scripts\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473226 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473282 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-dns-svc\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473318 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-run-httpd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473353 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473395 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-log-httpd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473444 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-config\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473472 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.473495 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.474102 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-fbfcr"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.475580 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrj8g\" (UniqueName: \"kubernetes.io/projected/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-kube-api-access-jrj8g\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.475626 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-config-data\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.477142 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-config\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.477155 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-dns-svc\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.477630 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.477974 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.482462 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.487921 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.516605 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrj8g\" (UniqueName: \"kubernetes.io/projected/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-kube-api-access-jrj8g\") pod \"dnsmasq-dns-f84976bdf-pwmxz\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: W1128 14:46:29.536369 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6da990a2_1382_4707_8311_5dd46e4f537c.slice/crio-87c5fbdb25d6896b406716797e4d96a1352c4fedf2c1c9d970c6af3330d7c2e5 WatchSource:0}: Error finding container 87c5fbdb25d6896b406716797e4d96a1352c4fedf2c1c9d970c6af3330d7c2e5: Status 404 returned error can't find the container with id 87c5fbdb25d6896b406716797e4d96a1352c4fedf2c1c9d970c6af3330d7c2e5 Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.576807 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb2tx\" (UniqueName: \"kubernetes.io/projected/fc32c327-72f5-444d-886e-b85a04ae31ba-kube-api-access-bb2tx\") pod \"fc32c327-72f5-444d-886e-b85a04ae31ba\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.583710 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-config\") pod \"fc32c327-72f5-444d-886e-b85a04ae31ba\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.583774 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-nb\") pod \"fc32c327-72f5-444d-886e-b85a04ae31ba\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.583816 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-sb\") pod \"fc32c327-72f5-444d-886e-b85a04ae31ba\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.583884 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-dns-svc\") pod \"fc32c327-72f5-444d-886e-b85a04ae31ba\" (UID: \"fc32c327-72f5-444d-886e-b85a04ae31ba\") " Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584240 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584305 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-config-data\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584399 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx8qd\" (UniqueName: \"kubernetes.io/projected/49a012fe-0369-4c46-89e8-0c0062f46c33-kube-api-access-gx8qd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584423 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-scripts\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584442 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584513 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-run-httpd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.584601 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-log-httpd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.585143 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-log-httpd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.590765 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-run-httpd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.595344 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-scripts\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.596789 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-config-data\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.597474 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.603541 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc32c327-72f5-444d-886e-b85a04ae31ba-kube-api-access-bb2tx" (OuterVolumeSpecName: "kube-api-access-bb2tx") pod "fc32c327-72f5-444d-886e-b85a04ae31ba" (UID: "fc32c327-72f5-444d-886e-b85a04ae31ba"). InnerVolumeSpecName "kube-api-access-bb2tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.606228 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.612145 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx8qd\" (UniqueName: \"kubernetes.io/projected/49a012fe-0369-4c46-89e8-0c0062f46c33-kube-api-access-gx8qd\") pod \"ceilometer-0\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.624182 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc32c327-72f5-444d-886e-b85a04ae31ba" (UID: "fc32c327-72f5-444d-886e-b85a04ae31ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.637399 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc32c327-72f5-444d-886e-b85a04ae31ba" (UID: "fc32c327-72f5-444d-886e-b85a04ae31ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.638837 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-config" (OuterVolumeSpecName: "config") pod "fc32c327-72f5-444d-886e-b85a04ae31ba" (UID: "fc32c327-72f5-444d-886e-b85a04ae31ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.641304 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.656214 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.670857 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc32c327-72f5-444d-886e-b85a04ae31ba" (UID: "fc32c327-72f5-444d-886e-b85a04ae31ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.687154 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb2tx\" (UniqueName: \"kubernetes.io/projected/fc32c327-72f5-444d-886e-b85a04ae31ba-kube-api-access-bb2tx\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.687193 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.687206 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.687217 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.687230 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc32c327-72f5-444d-886e-b85a04ae31ba-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.692295 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9lhpb"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.792024 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:29 crc kubenswrapper[4817]: E1128 14:46:29.792909 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc32c327-72f5-444d-886e-b85a04ae31ba" containerName="init" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.792925 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc32c327-72f5-444d-886e-b85a04ae31ba" containerName="init" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.793133 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc32c327-72f5-444d-886e-b85a04ae31ba" containerName="init" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.811101 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.811256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.816677 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wtq97" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.816897 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.819182 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.834451 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.836970 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.843814 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.855980 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892114 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892176 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-logs\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892249 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892268 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892284 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-config-data\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892303 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892337 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wg75\" (UniqueName: \"kubernetes.io/projected/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-kube-api-access-8wg75\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.892354 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.903872 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g85ld\" (UniqueName: \"kubernetes.io/projected/787427bb-4283-4537-af0a-ec8cb6072ea3-kube-api-access-g85ld\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.904058 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.904104 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-logs\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.904174 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.904220 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:29 crc kubenswrapper[4817]: I1128 14:46:29.904239 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-scripts\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005604 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-config-data\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005645 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005681 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wg75\" (UniqueName: \"kubernetes.io/projected/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-kube-api-access-8wg75\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005697 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005749 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g85ld\" (UniqueName: \"kubernetes.io/projected/787427bb-4283-4537-af0a-ec8cb6072ea3-kube-api-access-g85ld\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005766 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005784 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-logs\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005811 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005836 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005868 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-scripts\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005934 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.005976 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-logs\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.006008 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.006026 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.006589 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.006889 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.007225 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-logs\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.007712 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.011401 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-logs\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.011671 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.026198 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.033315 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.049902 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-jbnlp"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.115097 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-scripts\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.115652 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wg75\" (UniqueName: \"kubernetes.io/projected/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-kube-api-access-8wg75\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.116453 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-config-data\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.116501 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.117159 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g85ld\" (UniqueName: \"kubernetes.io/projected/787427bb-4283-4537-af0a-ec8cb6072ea3-kube-api-access-g85ld\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.117510 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.142136 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.159386 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fbfcr" event={"ID":"6da990a2-1382-4707-8311-5dd46e4f537c","Type":"ContainerStarted","Data":"87c5fbdb25d6896b406716797e4d96a1352c4fedf2c1c9d970c6af3330d7c2e5"} Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.161378 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9lhpb" event={"ID":"71e2f1db-1c92-4c41-9ef1-92fc41189f43","Type":"ContainerStarted","Data":"e358356b08c404ad7507596ce637df663db400b2cf0eea88c182a6169453e733"} Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.163298 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" event={"ID":"fc32c327-72f5-444d-886e-b85a04ae31ba","Type":"ContainerDied","Data":"554c471a9730bf399e88a01f271691d5ed1ddb65c2ad604efda288a1dc037560"} Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.163429 4817 scope.go:117] "RemoveContainer" containerID="e994d4d685f639d0fb47040768a2611c6c86b975dec8e8284382c36424c1189d" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.163560 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.163784 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-ps576" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.190418 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fflrd"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.218492 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"2228aaaf64eb2ec85cf7fc30aaed71088a168fb582ee3c0c122516f2598bf9a5"} Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.218774 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"42276c9f-fac8-4794-a69c-02178767e5b1","Type":"ContainerStarted","Data":"91f084594b9b9e1253b9c5ac0e75c5816ea5270bb219306b7db3cf9892b33c52"} Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.246381 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-ps576"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.253986 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-ps576"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.268545 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=47.2514671 podStartE2EDuration="58.268520869s" podCreationTimestamp="2025-11-28 14:45:32 +0000 UTC" firstStartedPulling="2025-11-28 14:46:12.999837408 +0000 UTC m=+995.587815674" lastFinishedPulling="2025-11-28 14:46:24.016891177 +0000 UTC m=+1006.604869443" observedRunningTime="2025-11-28 14:46:30.251804812 +0000 UTC m=+1012.839783078" watchObservedRunningTime="2025-11-28 14:46:30.268520869 +0000 UTC m=+1012.856499135" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.360349 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5q2qk"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.368631 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mpzzx"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.445602 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.470942 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.538489 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gvbcw"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.556260 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-pwmxz"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.598683 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-pwmxz"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.642581 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-c2qft"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.644435 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.660038 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.703783 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-c2qft"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.713235 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.730683 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-config\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.730748 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8hf7\" (UniqueName: \"kubernetes.io/projected/67758bba-e1c5-4315-8944-b1d247f2f422-kube-api-access-x8hf7\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.730782 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.730798 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.730827 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.730861 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.833689 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-config\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.833830 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8hf7\" (UniqueName: \"kubernetes.io/projected/67758bba-e1c5-4315-8944-b1d247f2f422-kube-api-access-x8hf7\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.833865 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.833880 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.833906 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.833943 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.834789 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.834795 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-config\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.834868 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.835256 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.838289 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.868024 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8hf7\" (UniqueName: \"kubernetes.io/projected/67758bba-e1c5-4315-8944-b1d247f2f422-kube-api-access-x8hf7\") pod \"dnsmasq-dns-785d8bcb8c-c2qft\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:30 crc kubenswrapper[4817]: I1128 14:46:30.964948 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:31 crc kubenswrapper[4817]: W1128 14:46:31.290706 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd69ea08f_64fd_49ca_be54_7dfc35fd5556.slice/crio-57d589ebd097170289cd6518b64325c7169908e2a069aae36e39a7ba365a170f WatchSource:0}: Error finding container 57d589ebd097170289cd6518b64325c7169908e2a069aae36e39a7ba365a170f: Status 404 returned error can't find the container with id 57d589ebd097170289cd6518b64325c7169908e2a069aae36e39a7ba365a170f Nov 28 14:46:31 crc kubenswrapper[4817]: W1128 14:46:31.305956 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca79eb0e_3393_4730_ba3a_c56fc3c6623f.slice/crio-6090483594e822d0e62effb8ad101c82383a3ce6984418660e1ff7092d22fa9b WatchSource:0}: Error finding container 6090483594e822d0e62effb8ad101c82383a3ce6984418660e1ff7092d22fa9b: Status 404 returned error can't find the container with id 6090483594e822d0e62effb8ad101c82383a3ce6984418660e1ff7092d22fa9b Nov 28 14:46:31 crc kubenswrapper[4817]: W1128 14:46:31.317039 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5784cff_bc3a_40d6_bf71_8678ac3ad73b.slice/crio-f96f751bbca474769abaab9216a8c486121ef1a502308b8f1fa5a7f8022bdd83 WatchSource:0}: Error finding container f96f751bbca474769abaab9216a8c486121ef1a502308b8f1fa5a7f8022bdd83: Status 404 returned error can't find the container with id f96f751bbca474769abaab9216a8c486121ef1a502308b8f1fa5a7f8022bdd83 Nov 28 14:46:31 crc kubenswrapper[4817]: W1128 14:46:31.317521 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49a012fe_0369_4c46_89e8_0c0062f46c33.slice/crio-60c7413466656756c933ce65e921123036d121b753c2c3d9be6716cc778b7ce9 WatchSource:0}: Error finding container 60c7413466656756c933ce65e921123036d121b753c2c3d9be6716cc778b7ce9: Status 404 returned error can't find the container with id 60c7413466656756c933ce65e921123036d121b753c2c3d9be6716cc778b7ce9 Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:31.750614 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc32c327-72f5-444d-886e-b85a04ae31ba" path="/var/lib/kubelet/pods/fc32c327-72f5-444d-886e-b85a04ae31ba/volumes" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:31.922870 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-c2qft"] Nov 28 14:46:33 crc kubenswrapper[4817]: W1128 14:46:31.923679 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67758bba_e1c5_4315_8944_b1d247f2f422.slice/crio-00ba7fbcb805553609bf843c2ee7a1af07758ab671cac5acb5ea19c5fff460db WatchSource:0}: Error finding container 00ba7fbcb805553609bf843c2ee7a1af07758ab671cac5acb5ea19c5fff460db: Status 404 returned error can't find the container with id 00ba7fbcb805553609bf843c2ee7a1af07758ab671cac5acb5ea19c5fff460db Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.010655 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.120209 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:33 crc kubenswrapper[4817]: W1128 14:46:32.136911 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod787427bb_4283_4537_af0a_ec8cb6072ea3.slice/crio-8a6a59a9fcddd2f31c836c831fbb9388e7cd58746b29f081c0b676c6587d91a0 WatchSource:0}: Error finding container 8a6a59a9fcddd2f31c836c831fbb9388e7cd58746b29f081c0b676c6587d91a0: Status 404 returned error can't find the container with id 8a6a59a9fcddd2f31c836c831fbb9388e7cd58746b29f081c0b676c6587d91a0 Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.236326 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" event={"ID":"d69ea08f-64fd-49ca-be54-7dfc35fd5556","Type":"ContainerStarted","Data":"57d589ebd097170289cd6518b64325c7169908e2a069aae36e39a7ba365a170f"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.237536 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvbcw" event={"ID":"3d26f30b-efdc-40ef-9ae5-ca78e79aba41","Type":"ContainerStarted","Data":"377e4e9171494829eb218364c92fa20a57c414340ee43107e92819698badf202"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.239064 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q2qk" event={"ID":"ca79eb0e-3393-4730-ba3a-c56fc3c6623f","Type":"ContainerStarted","Data":"6090483594e822d0e62effb8ad101c82383a3ce6984418660e1ff7092d22fa9b"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.240332 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerStarted","Data":"60c7413466656756c933ce65e921123036d121b753c2c3d9be6716cc778b7ce9"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.241546 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"787427bb-4283-4537-af0a-ec8cb6072ea3","Type":"ContainerStarted","Data":"8a6a59a9fcddd2f31c836c831fbb9388e7cd58746b29f081c0b676c6587d91a0"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.243659 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fflrd" event={"ID":"5d2d10b9-2269-4654-a979-cf2c9a8c65dc","Type":"ContainerStarted","Data":"fdbb165a75232b4190996ebb3f5eb44a8214a7bc0c3422327de8683659382259"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.246111 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mpzzx" event={"ID":"11e39ba1-49a5-4289-a913-10b87e82f4e0","Type":"ContainerStarted","Data":"7bd0f1b7af817d105f67194cd022d46acf73e3a5f38a63930ad24b85e9d4a03a"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.247251 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" event={"ID":"e5784cff-bc3a-40d6-bf71-8678ac3ad73b","Type":"ContainerStarted","Data":"f96f751bbca474769abaab9216a8c486121ef1a502308b8f1fa5a7f8022bdd83"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.248931 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"483fcf8d-7cb1-4d3e-9831-4d9111677ccf","Type":"ContainerStarted","Data":"e7e408eeed605d8f1485819dad939cfc4f8bd87afd68b75783ba9547d2b2d903"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.250201 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" event={"ID":"67758bba-e1c5-4315-8944-b1d247f2f422","Type":"ContainerStarted","Data":"00ba7fbcb805553609bf843c2ee7a1af07758ab671cac5acb5ea19c5fff460db"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.505761 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:32.645994 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.215555 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gk926"] Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.218235 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.231259 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gk926"] Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.263017 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"483fcf8d-7cb1-4d3e-9831-4d9111677ccf","Type":"ContainerStarted","Data":"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.264527 4817 generic.go:334] "Generic (PLEG): container finished" podID="d69ea08f-64fd-49ca-be54-7dfc35fd5556" containerID="0a06f5df2457e0316986ac9514f24bf51761dd126a277ad5ec17bb007c723ebe" exitCode=0 Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.264581 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" event={"ID":"d69ea08f-64fd-49ca-be54-7dfc35fd5556","Type":"ContainerDied","Data":"0a06f5df2457e0316986ac9514f24bf51761dd126a277ad5ec17bb007c723ebe"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.270282 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9lhpb" event={"ID":"71e2f1db-1c92-4c41-9ef1-92fc41189f43","Type":"ContainerStarted","Data":"a5fd050353186666463bdba9e4211db37581f9eb567c83012d2066140082892a"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.281711 4817 generic.go:334] "Generic (PLEG): container finished" podID="67758bba-e1c5-4315-8944-b1d247f2f422" containerID="367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2" exitCode=0 Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.281802 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" event={"ID":"67758bba-e1c5-4315-8944-b1d247f2f422","Type":"ContainerDied","Data":"367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.283463 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-utilities\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.283542 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-catalog-content\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.283576 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt2rr\" (UniqueName: \"kubernetes.io/projected/bf21de8d-ee94-4413-b5ef-203c656d1a93-kube-api-access-kt2rr\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.340998 4817 generic.go:334] "Generic (PLEG): container finished" podID="e5784cff-bc3a-40d6-bf71-8678ac3ad73b" containerID="c1b1939dbe933448698cceccf9d7634cdd765e8197cb11c4f4f08f19b9f0b7b3" exitCode=0 Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.341094 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" event={"ID":"e5784cff-bc3a-40d6-bf71-8678ac3ad73b","Type":"ContainerDied","Data":"c1b1939dbe933448698cceccf9d7634cdd765e8197cb11c4f4f08f19b9f0b7b3"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.365321 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"787427bb-4283-4537-af0a-ec8cb6072ea3","Type":"ContainerStarted","Data":"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.368373 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9lhpb" podStartSLOduration=5.368353342 podStartE2EDuration="5.368353342s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:46:33.365534742 +0000 UTC m=+1015.953513008" watchObservedRunningTime="2025-11-28 14:46:33.368353342 +0000 UTC m=+1015.956331608" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.394790 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-catalog-content\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.394861 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt2rr\" (UniqueName: \"kubernetes.io/projected/bf21de8d-ee94-4413-b5ef-203c656d1a93-kube-api-access-kt2rr\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.395061 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-utilities\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.404202 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-utilities\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.404741 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-catalog-content\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.422769 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fflrd" event={"ID":"5d2d10b9-2269-4654-a979-cf2c9a8c65dc","Type":"ContainerStarted","Data":"a9dce10085e4a2031130e9f98ffac108e9faf70fde37937df7241bc8a088d87d"} Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.465494 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt2rr\" (UniqueName: \"kubernetes.io/projected/bf21de8d-ee94-4413-b5ef-203c656d1a93-kube-api-access-kt2rr\") pod \"redhat-operators-gk926\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.560269 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-fflrd" podStartSLOduration=5.560250483 podStartE2EDuration="5.560250483s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:46:33.531167278 +0000 UTC m=+1016.119145544" watchObservedRunningTime="2025-11-28 14:46:33.560250483 +0000 UTC m=+1016.148228749" Nov 28 14:46:33 crc kubenswrapper[4817]: I1128 14:46:33.572265 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.043147 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.320419 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.323564 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.385342 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gk926"] Nov 28 14:46:34 crc kubenswrapper[4817]: W1128 14:46:34.395596 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf21de8d_ee94_4413_b5ef_203c656d1a93.slice/crio-72953d6dbe7313e15c3297aee28d2722b89bf43082793b73cbfc791738b82f9f WatchSource:0}: Error finding container 72953d6dbe7313e15c3297aee28d2722b89bf43082793b73cbfc791738b82f9f: Status 404 returned error can't find the container with id 72953d6dbe7313e15c3297aee28d2722b89bf43082793b73cbfc791738b82f9f Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433518 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-dns-svc\") pod \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433584 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-nb\") pod \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433658 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-nb\") pod \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433678 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-config\") pod \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433775 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-config\") pod \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433843 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrj8g\" (UniqueName: \"kubernetes.io/projected/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-kube-api-access-jrj8g\") pod \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433899 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-sb\") pod \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433929 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmmmn\" (UniqueName: \"kubernetes.io/projected/d69ea08f-64fd-49ca-be54-7dfc35fd5556-kube-api-access-cmmmn\") pod \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\" (UID: \"d69ea08f-64fd-49ca-be54-7dfc35fd5556\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433964 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-sb\") pod \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.433991 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-dns-svc\") pod \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\" (UID: \"e5784cff-bc3a-40d6-bf71-8678ac3ad73b\") " Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.494258 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-kube-api-access-jrj8g" (OuterVolumeSpecName: "kube-api-access-jrj8g") pod "e5784cff-bc3a-40d6-bf71-8678ac3ad73b" (UID: "e5784cff-bc3a-40d6-bf71-8678ac3ad73b"). InnerVolumeSpecName "kube-api-access-jrj8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.498159 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"787427bb-4283-4537-af0a-ec8cb6072ea3","Type":"ContainerStarted","Data":"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a"} Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.498320 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-log" containerID="cri-o://eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c" gracePeriod=30 Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.498414 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-httpd" containerID="cri-o://2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a" gracePeriod=30 Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.509787 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d69ea08f-64fd-49ca-be54-7dfc35fd5556-kube-api-access-cmmmn" (OuterVolumeSpecName: "kube-api-access-cmmmn") pod "d69ea08f-64fd-49ca-be54-7dfc35fd5556" (UID: "d69ea08f-64fd-49ca-be54-7dfc35fd5556"). InnerVolumeSpecName "kube-api-access-cmmmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.511256 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"483fcf8d-7cb1-4d3e-9831-4d9111677ccf","Type":"ContainerStarted","Data":"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d"} Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.511423 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-log" containerID="cri-o://242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c" gracePeriod=30 Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.511475 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-httpd" containerID="cri-o://b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d" gracePeriod=30 Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.514948 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e5784cff-bc3a-40d6-bf71-8678ac3ad73b" (UID: "e5784cff-bc3a-40d6-bf71-8678ac3ad73b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.519129 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerStarted","Data":"72953d6dbe7313e15c3297aee28d2722b89bf43082793b73cbfc791738b82f9f"} Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.520375 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5784cff-bc3a-40d6-bf71-8678ac3ad73b" (UID: "e5784cff-bc3a-40d6-bf71-8678ac3ad73b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.525823 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e5784cff-bc3a-40d6-bf71-8678ac3ad73b" (UID: "e5784cff-bc3a-40d6-bf71-8678ac3ad73b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.526272 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d69ea08f-64fd-49ca-be54-7dfc35fd5556" (UID: "d69ea08f-64fd-49ca-be54-7dfc35fd5556"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.526532 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-config" (OuterVolumeSpecName: "config") pod "d69ea08f-64fd-49ca-be54-7dfc35fd5556" (UID: "d69ea08f-64fd-49ca-be54-7dfc35fd5556"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.529875 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.529845498 podStartE2EDuration="6.529845498s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:46:34.52630889 +0000 UTC m=+1017.114287156" watchObservedRunningTime="2025-11-28 14:46:34.529845498 +0000 UTC m=+1017.117823854" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.537814 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" event={"ID":"d69ea08f-64fd-49ca-be54-7dfc35fd5556","Type":"ContainerDied","Data":"57d589ebd097170289cd6518b64325c7169908e2a069aae36e39a7ba365a170f"} Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.537876 4817 scope.go:117] "RemoveContainer" containerID="0a06f5df2457e0316986ac9514f24bf51761dd126a277ad5ec17bb007c723ebe" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.537888 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-jbnlp" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540157 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540180 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540191 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrj8g\" (UniqueName: \"kubernetes.io/projected/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-kube-api-access-jrj8g\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540201 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmmmn\" (UniqueName: \"kubernetes.io/projected/d69ea08f-64fd-49ca-be54-7dfc35fd5556-kube-api-access-cmmmn\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540210 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540218 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.540226 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.554195 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" event={"ID":"67758bba-e1c5-4315-8944-b1d247f2f422","Type":"ContainerStarted","Data":"a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c"} Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.554280 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d69ea08f-64fd-49ca-be54-7dfc35fd5556" (UID: "d69ea08f-64fd-49ca-be54-7dfc35fd5556"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.554888 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.567259 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.567629 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-pwmxz" event={"ID":"e5784cff-bc3a-40d6-bf71-8678ac3ad73b","Type":"ContainerDied","Data":"f96f751bbca474769abaab9216a8c486121ef1a502308b8f1fa5a7f8022bdd83"} Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.575872 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d69ea08f-64fd-49ca-be54-7dfc35fd5556" (UID: "d69ea08f-64fd-49ca-be54-7dfc35fd5556"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.579227 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-config" (OuterVolumeSpecName: "config") pod "e5784cff-bc3a-40d6-bf71-8678ac3ad73b" (UID: "e5784cff-bc3a-40d6-bf71-8678ac3ad73b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.580114 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.580084859 podStartE2EDuration="6.580084859s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:46:34.553699452 +0000 UTC m=+1017.141677718" watchObservedRunningTime="2025-11-28 14:46:34.580084859 +0000 UTC m=+1017.168063125" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.588304 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" podStartSLOduration=4.587874843 podStartE2EDuration="4.587874843s" podCreationTimestamp="2025-11-28 14:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:46:34.578812068 +0000 UTC m=+1017.166790334" watchObservedRunningTime="2025-11-28 14:46:34.587874843 +0000 UTC m=+1017.175853129" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.652982 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5784cff-bc3a-40d6-bf71-8678ac3ad73b-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.653016 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.653026 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d69ea08f-64fd-49ca-be54-7dfc35fd5556-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:34 crc kubenswrapper[4817]: I1128 14:46:34.670872 4817 scope.go:117] "RemoveContainer" containerID="c1b1939dbe933448698cceccf9d7634cdd765e8197cb11c4f4f08f19b9f0b7b3" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.008957 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-jbnlp"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.019051 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-jbnlp"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.038602 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-pwmxz"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.048454 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-pwmxz"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.281102 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376264 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wg75\" (UniqueName: \"kubernetes.io/projected/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-kube-api-access-8wg75\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376305 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376349 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-logs\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376373 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-scripts\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376450 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-combined-ca-bundle\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376503 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-config-data\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.376539 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-httpd-run\") pod \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\" (UID: \"483fcf8d-7cb1-4d3e-9831-4d9111677ccf\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.377007 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-logs" (OuterVolumeSpecName: "logs") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.377113 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.383884 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-kube-api-access-8wg75" (OuterVolumeSpecName: "kube-api-access-8wg75") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "kube-api-access-8wg75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.383908 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-scripts" (OuterVolumeSpecName: "scripts") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.384190 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.406373 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.469810 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-config-data" (OuterVolumeSpecName: "config-data") pod "483fcf8d-7cb1-4d3e-9831-4d9111677ccf" (UID: "483fcf8d-7cb1-4d3e-9831-4d9111677ccf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479129 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479166 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479176 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479186 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479195 4817 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479204 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wg75\" (UniqueName: \"kubernetes.io/projected/483fcf8d-7cb1-4d3e-9831-4d9111677ccf-kube-api-access-8wg75\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.479234 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.500477 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.584046 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.617909 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.626874 4817 generic.go:334] "Generic (PLEG): container finished" podID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerID="2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a" exitCode=0 Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.626922 4817 generic.go:334] "Generic (PLEG): container finished" podID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerID="eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c" exitCode=143 Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.626971 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"787427bb-4283-4537-af0a-ec8cb6072ea3","Type":"ContainerDied","Data":"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.627003 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"787427bb-4283-4537-af0a-ec8cb6072ea3","Type":"ContainerDied","Data":"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.627017 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"787427bb-4283-4537-af0a-ec8cb6072ea3","Type":"ContainerDied","Data":"8a6a59a9fcddd2f31c836c831fbb9388e7cd58746b29f081c0b676c6587d91a0"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.627035 4817 scope.go:117] "RemoveContainer" containerID="2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.630873 4817 generic.go:334] "Generic (PLEG): container finished" podID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerID="b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d" exitCode=143 Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.630894 4817 generic.go:334] "Generic (PLEG): container finished" podID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerID="242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c" exitCode=143 Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.630928 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"483fcf8d-7cb1-4d3e-9831-4d9111677ccf","Type":"ContainerDied","Data":"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.630945 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"483fcf8d-7cb1-4d3e-9831-4d9111677ccf","Type":"ContainerDied","Data":"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.630954 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"483fcf8d-7cb1-4d3e-9831-4d9111677ccf","Type":"ContainerDied","Data":"e7e408eeed605d8f1485819dad939cfc4f8bd87afd68b75783ba9547d2b2d903"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.631008 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.665948 4817 generic.go:334] "Generic (PLEG): container finished" podID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerID="a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea" exitCode=0 Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.666310 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerDied","Data":"a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea"} Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.685837 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-combined-ca-bundle\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.685945 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-scripts\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.686134 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.686185 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g85ld\" (UniqueName: \"kubernetes.io/projected/787427bb-4283-4537-af0a-ec8cb6072ea3-kube-api-access-g85ld\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.686229 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-logs\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.686272 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-httpd-run\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.686355 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-config-data\") pod \"787427bb-4283-4537-af0a-ec8cb6072ea3\" (UID: \"787427bb-4283-4537-af0a-ec8cb6072ea3\") " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.688126 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-logs" (OuterVolumeSpecName: "logs") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.688335 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.735037 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/787427bb-4283-4537-af0a-ec8cb6072ea3-kube-api-access-g85ld" (OuterVolumeSpecName: "kube-api-access-g85ld") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "kube-api-access-g85ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.758936 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-scripts" (OuterVolumeSpecName: "scripts") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.773037 4817 scope.go:117] "RemoveContainer" containerID="eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.773519 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.789774 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.789799 4817 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/787427bb-4283-4537-af0a-ec8cb6072ea3-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.789807 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.789825 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.789835 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g85ld\" (UniqueName: \"kubernetes.io/projected/787427bb-4283-4537-af0a-ec8cb6072ea3-kube-api-access-g85ld\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.791001 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d69ea08f-64fd-49ca-be54-7dfc35fd5556" path="/var/lib/kubelet/pods/d69ea08f-64fd-49ca-be54-7dfc35fd5556/volumes" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.791666 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5784cff-bc3a-40d6-bf71-8678ac3ad73b" path="/var/lib/kubelet/pods/e5784cff-bc3a-40d6-bf71-8678ac3ad73b/volumes" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.792053 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.814764 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.877773 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.877828 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.877846 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:35 crc kubenswrapper[4817]: E1128 14:46:35.878360 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5784cff-bc3a-40d6-bf71-8678ac3ad73b" containerName="init" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878374 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5784cff-bc3a-40d6-bf71-8678ac3ad73b" containerName="init" Nov 28 14:46:35 crc kubenswrapper[4817]: E1128 14:46:35.878390 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-log" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878415 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-log" Nov 28 14:46:35 crc kubenswrapper[4817]: E1128 14:46:35.878424 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-log" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878431 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-log" Nov 28 14:46:35 crc kubenswrapper[4817]: E1128 14:46:35.878444 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-httpd" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878449 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-httpd" Nov 28 14:46:35 crc kubenswrapper[4817]: E1128 14:46:35.878467 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69ea08f-64fd-49ca-be54-7dfc35fd5556" containerName="init" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878492 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69ea08f-64fd-49ca-be54-7dfc35fd5556" containerName="init" Nov 28 14:46:35 crc kubenswrapper[4817]: E1128 14:46:35.878505 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-httpd" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878511 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-httpd" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878736 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-httpd" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878761 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-log" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878775 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" containerName="glance-log" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878822 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" containerName="glance-httpd" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878835 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d69ea08f-64fd-49ca-be54-7dfc35fd5556" containerName="init" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.878845 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5784cff-bc3a-40d6-bf71-8678ac3ad73b" containerName="init" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.879996 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.880109 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.888617 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.891275 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.891297 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.920775 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-config-data" (OuterVolumeSpecName: "config-data") pod "787427bb-4283-4537-af0a-ec8cb6072ea3" (UID: "787427bb-4283-4537-af0a-ec8cb6072ea3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.992576 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.992621 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-config-data\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.992957 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.993100 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.993258 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-logs\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.993336 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q2jf\" (UniqueName: \"kubernetes.io/projected/f2978b05-83f7-4b0b-a9e7-a3748ff77920-kube-api-access-4q2jf\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.993408 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-scripts\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:35 crc kubenswrapper[4817]: I1128 14:46:35.993529 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787427bb-4283-4537-af0a-ec8cb6072ea3-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094600 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-logs\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094657 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q2jf\" (UniqueName: \"kubernetes.io/projected/f2978b05-83f7-4b0b-a9e7-a3748ff77920-kube-api-access-4q2jf\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094683 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-scripts\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094737 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094763 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-config-data\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094828 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.094870 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.096124 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.096374 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-logs\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.096766 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.100317 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.101036 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-scripts\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.116263 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q2jf\" (UniqueName: \"kubernetes.io/projected/f2978b05-83f7-4b0b-a9e7-a3748ff77920-kube-api-access-4q2jf\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.117952 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-config-data\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.129800 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.201628 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.386417 4817 scope.go:117] "RemoveContainer" containerID="2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a" Nov 28 14:46:36 crc kubenswrapper[4817]: E1128 14:46:36.386824 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a\": container with ID starting with 2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a not found: ID does not exist" containerID="2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.386882 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a"} err="failed to get container status \"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a\": rpc error: code = NotFound desc = could not find container \"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a\": container with ID starting with 2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.386907 4817 scope.go:117] "RemoveContainer" containerID="eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c" Nov 28 14:46:36 crc kubenswrapper[4817]: E1128 14:46:36.387441 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c\": container with ID starting with eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c not found: ID does not exist" containerID="eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.387505 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c"} err="failed to get container status \"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c\": rpc error: code = NotFound desc = could not find container \"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c\": container with ID starting with eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.387550 4817 scope.go:117] "RemoveContainer" containerID="2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.388044 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a"} err="failed to get container status \"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a\": rpc error: code = NotFound desc = could not find container \"2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a\": container with ID starting with 2c78b85661cbdd41dc541fbf7b05d6a772148d9f3c6663c338f6f9e9cc16ee2a not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.388098 4817 scope.go:117] "RemoveContainer" containerID="eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.389263 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c"} err="failed to get container status \"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c\": rpc error: code = NotFound desc = could not find container \"eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c\": container with ID starting with eb8078a8babf15c4e351a244ea051bb11156ef9f57177f0c808fbb465638a99c not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.389291 4817 scope.go:117] "RemoveContainer" containerID="b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.445738 4817 scope.go:117] "RemoveContainer" containerID="242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.487048 4817 scope.go:117] "RemoveContainer" containerID="b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d" Nov 28 14:46:36 crc kubenswrapper[4817]: E1128 14:46:36.487634 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d\": container with ID starting with b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d not found: ID does not exist" containerID="b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.487764 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d"} err="failed to get container status \"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d\": rpc error: code = NotFound desc = could not find container \"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d\": container with ID starting with b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.487835 4817 scope.go:117] "RemoveContainer" containerID="242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c" Nov 28 14:46:36 crc kubenswrapper[4817]: E1128 14:46:36.488224 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c\": container with ID starting with 242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c not found: ID does not exist" containerID="242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.488264 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c"} err="failed to get container status \"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c\": rpc error: code = NotFound desc = could not find container \"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c\": container with ID starting with 242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.488288 4817 scope.go:117] "RemoveContainer" containerID="b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.488655 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d"} err="failed to get container status \"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d\": rpc error: code = NotFound desc = could not find container \"b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d\": container with ID starting with b97a4ad6795b983b9c130ccda7cf34a8bbdf88d73677bd073e7a4983dbb2171d not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.488765 4817 scope.go:117] "RemoveContainer" containerID="242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.489333 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c"} err="failed to get container status \"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c\": rpc error: code = NotFound desc = could not find container \"242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c\": container with ID starting with 242ecfebf027f94bc9a9697ebe81ea5db0f4974f81cad78ace31e7710cdbca6c not found: ID does not exist" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.741397 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.764399 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.886157 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.904955 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.920843 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.922304 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.929113 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.929338 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 14:46:36 crc kubenswrapper[4817]: I1128 14:46:36.931350 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.047697 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.047779 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.047839 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.047873 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgtd4\" (UniqueName: \"kubernetes.io/projected/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-kube-api-access-kgtd4\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.047906 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-logs\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.047932 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.048001 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.048178 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.107335 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:46:37 crc kubenswrapper[4817]: W1128 14:46:37.145665 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2978b05_83f7_4b0b_a9e7_a3748ff77920.slice/crio-a74264d7f9417ecf4187f46f3224043ba3b54bb1df007f5beff3f1d78f8d6dd7 WatchSource:0}: Error finding container a74264d7f9417ecf4187f46f3224043ba3b54bb1df007f5beff3f1d78f8d6dd7: Status 404 returned error can't find the container with id a74264d7f9417ecf4187f46f3224043ba3b54bb1df007f5beff3f1d78f8d6dd7 Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162070 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162136 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162166 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162214 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162257 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgtd4\" (UniqueName: \"kubernetes.io/projected/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-kube-api-access-kgtd4\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162291 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-logs\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162315 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162362 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.162875 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.163852 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.166130 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-logs\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.173553 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.173553 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.173897 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.179563 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.182839 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgtd4\" (UniqueName: \"kubernetes.io/projected/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-kube-api-access-kgtd4\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.198713 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.256576 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.754615 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483fcf8d-7cb1-4d3e-9831-4d9111677ccf" path="/var/lib/kubelet/pods/483fcf8d-7cb1-4d3e-9831-4d9111677ccf/volumes" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.756429 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="787427bb-4283-4537-af0a-ec8cb6072ea3" path="/var/lib/kubelet/pods/787427bb-4283-4537-af0a-ec8cb6072ea3/volumes" Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.862947 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2978b05-83f7-4b0b-a9e7-a3748ff77920","Type":"ContainerStarted","Data":"a74264d7f9417ecf4187f46f3224043ba3b54bb1df007f5beff3f1d78f8d6dd7"} Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.863317 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:46:37 crc kubenswrapper[4817]: I1128 14:46:37.868001 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerStarted","Data":"38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13"} Nov 28 14:46:38 crc kubenswrapper[4817]: I1128 14:46:38.876452 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1","Type":"ContainerStarted","Data":"af321c9de28d682572ce61939bc206feb56547cfcc0fb2b598c534185238b284"} Nov 28 14:46:39 crc kubenswrapper[4817]: I1128 14:46:39.895016 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2978b05-83f7-4b0b-a9e7-a3748ff77920","Type":"ContainerStarted","Data":"07641ed66bfb821245092930ed01fe40782c7ebc2a3212ae2f14f1660dd590ff"} Nov 28 14:46:39 crc kubenswrapper[4817]: I1128 14:46:39.898175 4817 generic.go:334] "Generic (PLEG): container finished" podID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerID="38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13" exitCode=0 Nov 28 14:46:39 crc kubenswrapper[4817]: I1128 14:46:39.898215 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerDied","Data":"38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13"} Nov 28 14:46:40 crc kubenswrapper[4817]: I1128 14:46:40.965840 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:46:41 crc kubenswrapper[4817]: I1128 14:46:41.044182 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pr6vs"] Nov 28 14:46:41 crc kubenswrapper[4817]: I1128 14:46:41.046250 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-pr6vs" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" containerID="cri-o://3f32a9dec98e4bfd02513cbbcf1005b69f8f8798fb6b0d6c710039c661bcec02" gracePeriod=10 Nov 28 14:46:42 crc kubenswrapper[4817]: I1128 14:46:42.404429 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-pr6vs" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Nov 28 14:46:44 crc kubenswrapper[4817]: I1128 14:46:44.939359 4817 generic.go:334] "Generic (PLEG): container finished" podID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerID="3f32a9dec98e4bfd02513cbbcf1005b69f8f8798fb6b0d6c710039c661bcec02" exitCode=0 Nov 28 14:46:44 crc kubenswrapper[4817]: I1128 14:46:44.939447 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pr6vs" event={"ID":"6586dc95-7ca6-476a-a7d7-67e550cb19e9","Type":"ContainerDied","Data":"3f32a9dec98e4bfd02513cbbcf1005b69f8f8798fb6b0d6c710039c661bcec02"} Nov 28 14:46:45 crc kubenswrapper[4817]: I1128 14:46:45.949262 4817 generic.go:334] "Generic (PLEG): container finished" podID="71e2f1db-1c92-4c41-9ef1-92fc41189f43" containerID="a5fd050353186666463bdba9e4211db37581f9eb567c83012d2066140082892a" exitCode=0 Nov 28 14:46:45 crc kubenswrapper[4817]: I1128 14:46:45.949317 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9lhpb" event={"ID":"71e2f1db-1c92-4c41-9ef1-92fc41189f43","Type":"ContainerDied","Data":"a5fd050353186666463bdba9e4211db37581f9eb567c83012d2066140082892a"} Nov 28 14:46:50 crc kubenswrapper[4817]: E1128 14:46:50.342671 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Nov 28 14:46:50 crc kubenswrapper[4817]: E1128 14:46:50.343255 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dw4g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-fbfcr_openstack(6da990a2-1382-4707-8311-5dd46e4f537c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:46:50 crc kubenswrapper[4817]: E1128 14:46:50.344448 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-fbfcr" podUID="6da990a2-1382-4707-8311-5dd46e4f537c" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.598306 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wmlbp"] Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.600508 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.611127 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wmlbp"] Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.748194 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-utilities\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.748247 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-catalog-content\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.748322 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmfqf\" (UniqueName: \"kubernetes.io/projected/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-kube-api-access-nmfqf\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.851842 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmfqf\" (UniqueName: \"kubernetes.io/projected/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-kube-api-access-nmfqf\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.854115 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-utilities\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.854153 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-catalog-content\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.854349 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-utilities\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.854470 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-catalog-content\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.869649 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmfqf\" (UniqueName: \"kubernetes.io/projected/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-kube-api-access-nmfqf\") pod \"certified-operators-wmlbp\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.933299 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:46:50 crc kubenswrapper[4817]: I1128 14:46:50.999867 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1","Type":"ContainerStarted","Data":"57478e58c4952c9a190ce4ef2fc26d7345be350a1e244deee775c1d100af9f6a"} Nov 28 14:46:51 crc kubenswrapper[4817]: E1128 14:46:51.001012 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-fbfcr" podUID="6da990a2-1382-4707-8311-5dd46e4f537c" Nov 28 14:46:52 crc kubenswrapper[4817]: I1128 14:46:52.405469 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-pr6vs" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: i/o timeout" Nov 28 14:46:54 crc kubenswrapper[4817]: I1128 14:46:54.046640 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:46:54 crc kubenswrapper[4817]: I1128 14:46:54.047072 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:46:54 crc kubenswrapper[4817]: I1128 14:46:54.047114 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:46:54 crc kubenswrapper[4817]: I1128 14:46:54.047770 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f352602e89e72622a4635f595006f3c005b7dadf542d0db30e6fcfcc7b955afc"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:46:54 crc kubenswrapper[4817]: I1128 14:46:54.047816 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://f352602e89e72622a4635f595006f3c005b7dadf542d0db30e6fcfcc7b955afc" gracePeriod=600 Nov 28 14:46:55 crc kubenswrapper[4817]: I1128 14:46:55.042884 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="f352602e89e72622a4635f595006f3c005b7dadf542d0db30e6fcfcc7b955afc" exitCode=0 Nov 28 14:46:55 crc kubenswrapper[4817]: I1128 14:46:55.043044 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"f352602e89e72622a4635f595006f3c005b7dadf542d0db30e6fcfcc7b955afc"} Nov 28 14:46:55 crc kubenswrapper[4817]: I1128 14:46:55.043388 4817 scope.go:117] "RemoveContainer" containerID="e638df1b21b62724d5463d8dfc0e4379271e7a660caed8cd6069eeac89d71b58" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.184659 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.191007 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.275388 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-combined-ca-bundle\") pod \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.275453 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-credential-keys\") pod \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.275539 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-nb\") pod \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.275584 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-config-data\") pod \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.275918 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8g5c\" (UniqueName: \"kubernetes.io/projected/6586dc95-7ca6-476a-a7d7-67e550cb19e9-kube-api-access-r8g5c\") pod \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.275948 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvb5j\" (UniqueName: \"kubernetes.io/projected/71e2f1db-1c92-4c41-9ef1-92fc41189f43-kube-api-access-jvb5j\") pod \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.276004 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-sb\") pod \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.276055 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-fernet-keys\") pod \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.276088 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-config\") pod \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.276133 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-scripts\") pod \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\" (UID: \"71e2f1db-1c92-4c41-9ef1-92fc41189f43\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.276183 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-dns-svc\") pod \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\" (UID: \"6586dc95-7ca6-476a-a7d7-67e550cb19e9\") " Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.325058 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "71e2f1db-1c92-4c41-9ef1-92fc41189f43" (UID: "71e2f1db-1c92-4c41-9ef1-92fc41189f43"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.325088 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6586dc95-7ca6-476a-a7d7-67e550cb19e9-kube-api-access-r8g5c" (OuterVolumeSpecName: "kube-api-access-r8g5c") pod "6586dc95-7ca6-476a-a7d7-67e550cb19e9" (UID: "6586dc95-7ca6-476a-a7d7-67e550cb19e9"). InnerVolumeSpecName "kube-api-access-r8g5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.325122 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e2f1db-1c92-4c41-9ef1-92fc41189f43-kube-api-access-jvb5j" (OuterVolumeSpecName: "kube-api-access-jvb5j") pod "71e2f1db-1c92-4c41-9ef1-92fc41189f43" (UID: "71e2f1db-1c92-4c41-9ef1-92fc41189f43"). InnerVolumeSpecName "kube-api-access-jvb5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.325209 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "71e2f1db-1c92-4c41-9ef1-92fc41189f43" (UID: "71e2f1db-1c92-4c41-9ef1-92fc41189f43"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.364824 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-scripts" (OuterVolumeSpecName: "scripts") pod "71e2f1db-1c92-4c41-9ef1-92fc41189f43" (UID: "71e2f1db-1c92-4c41-9ef1-92fc41189f43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.374529 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71e2f1db-1c92-4c41-9ef1-92fc41189f43" (UID: "71e2f1db-1c92-4c41-9ef1-92fc41189f43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.377646 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8g5c\" (UniqueName: \"kubernetes.io/projected/6586dc95-7ca6-476a-a7d7-67e550cb19e9-kube-api-access-r8g5c\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.377668 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvb5j\" (UniqueName: \"kubernetes.io/projected/71e2f1db-1c92-4c41-9ef1-92fc41189f43-kube-api-access-jvb5j\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.377677 4817 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.377685 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.377694 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.377702 4817 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.408201 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-pr6vs" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: i/o timeout" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.408471 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.431566 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-config-data" (OuterVolumeSpecName: "config-data") pod "71e2f1db-1c92-4c41-9ef1-92fc41189f43" (UID: "71e2f1db-1c92-4c41-9ef1-92fc41189f43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.444517 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6586dc95-7ca6-476a-a7d7-67e550cb19e9" (UID: "6586dc95-7ca6-476a-a7d7-67e550cb19e9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.444975 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-config" (OuterVolumeSpecName: "config") pod "6586dc95-7ca6-476a-a7d7-67e550cb19e9" (UID: "6586dc95-7ca6-476a-a7d7-67e550cb19e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.456969 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6586dc95-7ca6-476a-a7d7-67e550cb19e9" (UID: "6586dc95-7ca6-476a-a7d7-67e550cb19e9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.459293 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6586dc95-7ca6-476a-a7d7-67e550cb19e9" (UID: "6586dc95-7ca6-476a-a7d7-67e550cb19e9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.479003 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.479039 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.479052 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.479065 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71e2f1db-1c92-4c41-9ef1-92fc41189f43-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:57 crc kubenswrapper[4817]: I1128 14:46:57.479077 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6586dc95-7ca6-476a-a7d7-67e550cb19e9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.073199 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9lhpb" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.073242 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9lhpb" event={"ID":"71e2f1db-1c92-4c41-9ef1-92fc41189f43","Type":"ContainerDied","Data":"e358356b08c404ad7507596ce637df663db400b2cf0eea88c182a6169453e733"} Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.073283 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e358356b08c404ad7507596ce637df663db400b2cf0eea88c182a6169453e733" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.075836 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-pr6vs" event={"ID":"6586dc95-7ca6-476a-a7d7-67e550cb19e9","Type":"ContainerDied","Data":"a2ad9a39b4f592c38c69da82daa18719dbe38fb58d98d0b02903c4f2dc08c957"} Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.076091 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-pr6vs" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.146291 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pr6vs"] Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.157698 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-pr6vs"] Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.294008 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9lhpb"] Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.302634 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9lhpb"] Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.389258 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-lgl42"] Nov 28 14:46:58 crc kubenswrapper[4817]: E1128 14:46:58.389632 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="init" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.389649 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="init" Nov 28 14:46:58 crc kubenswrapper[4817]: E1128 14:46:58.389672 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.389680 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" Nov 28 14:46:58 crc kubenswrapper[4817]: E1128 14:46:58.389697 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e2f1db-1c92-4c41-9ef1-92fc41189f43" containerName="keystone-bootstrap" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.389703 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e2f1db-1c92-4c41-9ef1-92fc41189f43" containerName="keystone-bootstrap" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.389880 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e2f1db-1c92-4c41-9ef1-92fc41189f43" containerName="keystone-bootstrap" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.389898 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" containerName="dnsmasq-dns" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.390445 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.395068 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.395122 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.395200 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-knk6m" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.395433 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.395629 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.412067 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lgl42"] Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.497235 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-config-data\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.497288 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7hgq\" (UniqueName: \"kubernetes.io/projected/4482a16d-c783-4609-9db4-4e9eb3cb3c23-kube-api-access-m7hgq\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.497312 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-combined-ca-bundle\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.497366 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-scripts\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.497386 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-credential-keys\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.497481 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-fernet-keys\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.598579 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-config-data\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.598623 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7hgq\" (UniqueName: \"kubernetes.io/projected/4482a16d-c783-4609-9db4-4e9eb3cb3c23-kube-api-access-m7hgq\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.598646 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-combined-ca-bundle\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.598703 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-scripts\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.598742 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-credential-keys\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.599031 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-fernet-keys\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.603767 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-config-data\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.604197 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-credential-keys\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.619243 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-scripts\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.619552 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-combined-ca-bundle\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.619580 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-fernet-keys\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.622083 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7hgq\" (UniqueName: \"kubernetes.io/projected/4482a16d-c783-4609-9db4-4e9eb3cb3c23-kube-api-access-m7hgq\") pod \"keystone-bootstrap-lgl42\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:58 crc kubenswrapper[4817]: I1128 14:46:58.714793 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:46:59 crc kubenswrapper[4817]: I1128 14:46:59.751934 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6586dc95-7ca6-476a-a7d7-67e550cb19e9" path="/var/lib/kubelet/pods/6586dc95-7ca6-476a-a7d7-67e550cb19e9/volumes" Nov 28 14:46:59 crc kubenswrapper[4817]: I1128 14:46:59.753767 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71e2f1db-1c92-4c41-9ef1-92fc41189f43" path="/var/lib/kubelet/pods/71e2f1db-1c92-4c41-9ef1-92fc41189f43/volumes" Nov 28 14:47:07 crc kubenswrapper[4817]: E1128 14:47:07.104934 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 28 14:47:07 crc kubenswrapper[4817]: E1128 14:47:07.105557 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hkg8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-mpzzx_openstack(11e39ba1-49a5-4289-a913-10b87e82f4e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:47:07 crc kubenswrapper[4817]: E1128 14:47:07.106806 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-mpzzx" podUID="11e39ba1-49a5-4289-a913-10b87e82f4e0" Nov 28 14:47:07 crc kubenswrapper[4817]: E1128 14:47:07.158092 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-mpzzx" podUID="11e39ba1-49a5-4289-a913-10b87e82f4e0" Nov 28 14:47:07 crc kubenswrapper[4817]: E1128 14:47:07.537584 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 28 14:47:07 crc kubenswrapper[4817]: E1128 14:47:07.537774 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n584h64ch597h668h59dh59h545h654hbbh57h65bh67ch8dhf7h647h66dh68dhc7h5c5h644h5b4h545h649h559h5bdh85h67fh9dhb7h559h548hd5q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gx8qd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(49a012fe-0369-4c46-89e8-0c0062f46c33): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:47:08 crc kubenswrapper[4817]: E1128 14:47:08.757985 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 28 14:47:08 crc kubenswrapper[4817]: E1128 14:47:08.758361 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pfctv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-5q2qk_openstack(ca79eb0e-3393-4730-ba3a-c56fc3c6623f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:47:08 crc kubenswrapper[4817]: E1128 14:47:08.759553 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-5q2qk" podUID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" Nov 28 14:47:08 crc kubenswrapper[4817]: I1128 14:47:08.764431 4817 scope.go:117] "RemoveContainer" containerID="3f32a9dec98e4bfd02513cbbcf1005b69f8f8798fb6b0d6c710039c661bcec02" Nov 28 14:47:08 crc kubenswrapper[4817]: I1128 14:47:08.970442 4817 scope.go:117] "RemoveContainer" containerID="5335d7c7f38b3ce4ca3b2fa4985cce99d73208bfc7870db008f21dbf080b7efa" Nov 28 14:47:09 crc kubenswrapper[4817]: I1128 14:47:09.186901 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerStarted","Data":"7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e"} Nov 28 14:47:09 crc kubenswrapper[4817]: I1128 14:47:09.190967 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"6233c40182fb7115b99ef7443b2744ab4c0eae4c9222a791b73fbfc8281a9e58"} Nov 28 14:47:09 crc kubenswrapper[4817]: E1128 14:47:09.192762 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-5q2qk" podUID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" Nov 28 14:47:09 crc kubenswrapper[4817]: I1128 14:47:09.212784 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gk926" podStartSLOduration=3.14482018 podStartE2EDuration="36.212746896s" podCreationTimestamp="2025-11-28 14:46:33 +0000 UTC" firstStartedPulling="2025-11-28 14:46:35.684962845 +0000 UTC m=+1018.272941111" lastFinishedPulling="2025-11-28 14:47:08.752889561 +0000 UTC m=+1051.340867827" observedRunningTime="2025-11-28 14:47:09.210042559 +0000 UTC m=+1051.798020825" watchObservedRunningTime="2025-11-28 14:47:09.212746896 +0000 UTC m=+1051.800725152" Nov 28 14:47:09 crc kubenswrapper[4817]: I1128 14:47:09.284184 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lgl42"] Nov 28 14:47:09 crc kubenswrapper[4817]: W1128 14:47:09.302899 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4482a16d_c783_4609_9db4_4e9eb3cb3c23.slice/crio-cde776bb251d20603b7aa46b3a69b0cc3d92f86b92a1b97b6cc0a62bd0075ed3 WatchSource:0}: Error finding container cde776bb251d20603b7aa46b3a69b0cc3d92f86b92a1b97b6cc0a62bd0075ed3: Status 404 returned error can't find the container with id cde776bb251d20603b7aa46b3a69b0cc3d92f86b92a1b97b6cc0a62bd0075ed3 Nov 28 14:47:09 crc kubenswrapper[4817]: I1128 14:47:09.396446 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wmlbp"] Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.203637 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2978b05-83f7-4b0b-a9e7-a3748ff77920","Type":"ContainerStarted","Data":"4ba7285c34eb49f50f152a75a9053ec4a3f6de6a3479acf887d3761d6784f2bf"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.203733 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-log" containerID="cri-o://07641ed66bfb821245092930ed01fe40782c7ebc2a3212ae2f14f1660dd590ff" gracePeriod=30 Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.203821 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-httpd" containerID="cri-o://4ba7285c34eb49f50f152a75a9053ec4a3f6de6a3479acf887d3761d6784f2bf" gracePeriod=30 Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.208193 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1","Type":"ContainerStarted","Data":"7d6ec80e36c3086c48828d66ba9659a186575524e70c5be9355705832797e3d6"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.210357 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lgl42" event={"ID":"4482a16d-c783-4609-9db4-4e9eb3cb3c23","Type":"ContainerStarted","Data":"3a4db1f3b8d1fbb685dbb132cae457f5c5e1c0600980e7fb24b0d52bea7cff61"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.210389 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lgl42" event={"ID":"4482a16d-c783-4609-9db4-4e9eb3cb3c23","Type":"ContainerStarted","Data":"cde776bb251d20603b7aa46b3a69b0cc3d92f86b92a1b97b6cc0a62bd0075ed3"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.211977 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fbfcr" event={"ID":"6da990a2-1382-4707-8311-5dd46e4f537c","Type":"ContainerStarted","Data":"aa36b76a97ce87e35668be6b4c9bb5840bf2a30d6a66424da3c73491b117c116"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.214023 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvbcw" event={"ID":"3d26f30b-efdc-40ef-9ae5-ca78e79aba41","Type":"ContainerStarted","Data":"1da30229a0fbcadc4d0e4522431651004dbe41e73c325d9cfbbf4234aaf9f99e"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.215559 4817 generic.go:334] "Generic (PLEG): container finished" podID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerID="224926ac3023b14675f7a87fa20d95ea669b40299c9b456a8fdf8fac8bb58bc6" exitCode=0 Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.215632 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmlbp" event={"ID":"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026","Type":"ContainerDied","Data":"224926ac3023b14675f7a87fa20d95ea669b40299c9b456a8fdf8fac8bb58bc6"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.215976 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmlbp" event={"ID":"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026","Type":"ContainerStarted","Data":"429ad04dd947fe6d2dbeebaa9bbf17b4d031d828ae7c6d00d076b3abc0c32659"} Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.242419 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=35.242395479 podStartE2EDuration="35.242395479s" podCreationTimestamp="2025-11-28 14:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:10.230555604 +0000 UTC m=+1052.818533860" watchObservedRunningTime="2025-11-28 14:47:10.242395479 +0000 UTC m=+1052.830373745" Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.252229 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-lgl42" podStartSLOduration=12.252208972 podStartE2EDuration="12.252208972s" podCreationTimestamp="2025-11-28 14:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:10.250686074 +0000 UTC m=+1052.838664350" watchObservedRunningTime="2025-11-28 14:47:10.252208972 +0000 UTC m=+1052.840187238" Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.336193 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=34.336177639 podStartE2EDuration="34.336177639s" podCreationTimestamp="2025-11-28 14:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:10.309357012 +0000 UTC m=+1052.897335288" watchObservedRunningTime="2025-11-28 14:47:10.336177639 +0000 UTC m=+1052.924155905" Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.337033 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gvbcw" podStartSLOduration=6.174785473 podStartE2EDuration="42.33702811s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="2025-11-28 14:46:31.341162541 +0000 UTC m=+1013.929140807" lastFinishedPulling="2025-11-28 14:47:07.503405178 +0000 UTC m=+1050.091383444" observedRunningTime="2025-11-28 14:47:10.336870146 +0000 UTC m=+1052.924848422" watchObservedRunningTime="2025-11-28 14:47:10.33702811 +0000 UTC m=+1052.925006376" Nov 28 14:47:10 crc kubenswrapper[4817]: I1128 14:47:10.359510 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-fbfcr" podStartSLOduration=2.8494465829999998 podStartE2EDuration="42.359485258s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="2025-11-28 14:46:29.539467456 +0000 UTC m=+1012.127445722" lastFinishedPulling="2025-11-28 14:47:09.049506111 +0000 UTC m=+1051.637484397" observedRunningTime="2025-11-28 14:47:10.351418027 +0000 UTC m=+1052.939396303" watchObservedRunningTime="2025-11-28 14:47:10.359485258 +0000 UTC m=+1052.947463524" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.226760 4817 generic.go:334] "Generic (PLEG): container finished" podID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerID="4ba7285c34eb49f50f152a75a9053ec4a3f6de6a3479acf887d3761d6784f2bf" exitCode=0 Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.227416 4817 generic.go:334] "Generic (PLEG): container finished" podID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerID="07641ed66bfb821245092930ed01fe40782c7ebc2a3212ae2f14f1660dd590ff" exitCode=143 Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.228224 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2978b05-83f7-4b0b-a9e7-a3748ff77920","Type":"ContainerDied","Data":"4ba7285c34eb49f50f152a75a9053ec4a3f6de6a3479acf887d3761d6784f2bf"} Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.228248 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2978b05-83f7-4b0b-a9e7-a3748ff77920","Type":"ContainerDied","Data":"07641ed66bfb821245092930ed01fe40782c7ebc2a3212ae2f14f1660dd590ff"} Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.438164 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.538580 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-logs\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.538671 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-scripts\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.538854 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q2jf\" (UniqueName: \"kubernetes.io/projected/f2978b05-83f7-4b0b-a9e7-a3748ff77920-kube-api-access-4q2jf\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.538888 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-httpd-run\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.538925 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-config-data\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.538969 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.539014 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-combined-ca-bundle\") pod \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\" (UID: \"f2978b05-83f7-4b0b-a9e7-a3748ff77920\") " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.539159 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-logs" (OuterVolumeSpecName: "logs") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.539344 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.539416 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.539431 4817 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2978b05-83f7-4b0b-a9e7-a3748ff77920-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.556689 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.557097 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2978b05-83f7-4b0b-a9e7-a3748ff77920-kube-api-access-4q2jf" (OuterVolumeSpecName: "kube-api-access-4q2jf") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "kube-api-access-4q2jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.574541 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-scripts" (OuterVolumeSpecName: "scripts") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.609451 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.628126 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-config-data" (OuterVolumeSpecName: "config-data") pod "f2978b05-83f7-4b0b-a9e7-a3748ff77920" (UID: "f2978b05-83f7-4b0b-a9e7-a3748ff77920"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.643612 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.643657 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q2jf\" (UniqueName: \"kubernetes.io/projected/f2978b05-83f7-4b0b-a9e7-a3748ff77920-kube-api-access-4q2jf\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.643669 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.643703 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.643746 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2978b05-83f7-4b0b-a9e7-a3748ff77920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.670222 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 14:47:11 crc kubenswrapper[4817]: I1128 14:47:11.744814 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.237500 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2978b05-83f7-4b0b-a9e7-a3748ff77920","Type":"ContainerDied","Data":"a74264d7f9417ecf4187f46f3224043ba3b54bb1df007f5beff3f1d78f8d6dd7"} Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.237926 4817 scope.go:117] "RemoveContainer" containerID="4ba7285c34eb49f50f152a75a9053ec4a3f6de6a3479acf887d3761d6784f2bf" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.237537 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.246767 4817 generic.go:334] "Generic (PLEG): container finished" podID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerID="23d817e3e29bb98b2041aefa13adf7217406c3e7c421863d97ec0f86a7b51d0c" exitCode=0 Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.246863 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmlbp" event={"ID":"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026","Type":"ContainerDied","Data":"23d817e3e29bb98b2041aefa13adf7217406c3e7c421863d97ec0f86a7b51d0c"} Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.248935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerStarted","Data":"9c043a562b41a41ac20a4fc97fc6c7620409f0425c95c6f82e3900204828bea3"} Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.264992 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.273763 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.301064 4817 scope.go:117] "RemoveContainer" containerID="07641ed66bfb821245092930ed01fe40782c7ebc2a3212ae2f14f1660dd590ff" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.315797 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:47:12 crc kubenswrapper[4817]: E1128 14:47:12.316564 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-httpd" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.316585 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-httpd" Nov 28 14:47:12 crc kubenswrapper[4817]: E1128 14:47:12.316604 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-log" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.316610 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-log" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.316929 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-log" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.316954 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" containerName="glance-httpd" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.328153 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.337769 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.359218 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.359302 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473114 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473242 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473286 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-logs\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473310 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473341 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfzpm\" (UniqueName: \"kubernetes.io/projected/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-kube-api-access-vfzpm\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473374 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473425 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.473458 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575268 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575362 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575384 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575427 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575482 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575506 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-logs\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575526 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.575936 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.576788 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-logs\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.576912 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.577068 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfzpm\" (UniqueName: \"kubernetes.io/projected/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-kube-api-access-vfzpm\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.592153 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.592259 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.594305 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.600206 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.600672 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfzpm\" (UniqueName: \"kubernetes.io/projected/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-kube-api-access-vfzpm\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.620279 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " pod="openstack/glance-default-external-api-0" Nov 28 14:47:12 crc kubenswrapper[4817]: I1128 14:47:12.692685 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:47:13 crc kubenswrapper[4817]: I1128 14:47:13.573749 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:47:13 crc kubenswrapper[4817]: I1128 14:47:13.574070 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:47:13 crc kubenswrapper[4817]: I1128 14:47:13.751438 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2978b05-83f7-4b0b-a9e7-a3748ff77920" path="/var/lib/kubelet/pods/f2978b05-83f7-4b0b-a9e7-a3748ff77920/volumes" Nov 28 14:47:14 crc kubenswrapper[4817]: I1128 14:47:14.625673 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gk926" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" probeResult="failure" output=< Nov 28 14:47:14 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 14:47:14 crc kubenswrapper[4817]: > Nov 28 14:47:15 crc kubenswrapper[4817]: I1128 14:47:15.693030 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:47:16 crc kubenswrapper[4817]: I1128 14:47:16.297403 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9f53839-25c4-4e8f-b9c7-1895ddff10d6","Type":"ContainerStarted","Data":"ce198414b18af0d6b69e8ba0585f8f86425dec46b7ebbdf3fab5e0aa387bdc35"} Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.257603 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.257971 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.300579 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.310419 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.312743 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9f53839-25c4-4e8f-b9c7-1895ddff10d6","Type":"ContainerStarted","Data":"6fa052b7b40112b5990ebbff5d9a4b45c427e0cab7a5e5ecd75d6779e0bb83da"} Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.312964 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:17 crc kubenswrapper[4817]: I1128 14:47:17.313036 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:18 crc kubenswrapper[4817]: I1128 14:47:18.339978 4817 generic.go:334] "Generic (PLEG): container finished" podID="3d26f30b-efdc-40ef-9ae5-ca78e79aba41" containerID="1da30229a0fbcadc4d0e4522431651004dbe41e73c325d9cfbbf4234aaf9f99e" exitCode=0 Nov 28 14:47:18 crc kubenswrapper[4817]: I1128 14:47:18.340068 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvbcw" event={"ID":"3d26f30b-efdc-40ef-9ae5-ca78e79aba41","Type":"ContainerDied","Data":"1da30229a0fbcadc4d0e4522431651004dbe41e73c325d9cfbbf4234aaf9f99e"} Nov 28 14:47:18 crc kubenswrapper[4817]: I1128 14:47:18.347456 4817 generic.go:334] "Generic (PLEG): container finished" podID="4482a16d-c783-4609-9db4-4e9eb3cb3c23" containerID="3a4db1f3b8d1fbb685dbb132cae457f5c5e1c0600980e7fb24b0d52bea7cff61" exitCode=0 Nov 28 14:47:18 crc kubenswrapper[4817]: I1128 14:47:18.347517 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lgl42" event={"ID":"4482a16d-c783-4609-9db4-4e9eb3cb3c23","Type":"ContainerDied","Data":"3a4db1f3b8d1fbb685dbb132cae457f5c5e1c0600980e7fb24b0d52bea7cff61"} Nov 28 14:47:19 crc kubenswrapper[4817]: I1128 14:47:19.414026 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:19 crc kubenswrapper[4817]: I1128 14:47:19.414477 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.296001 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvbcw" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.308162 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.366816 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-fernet-keys\") pod \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.366945 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7hgq\" (UniqueName: \"kubernetes.io/projected/4482a16d-c783-4609-9db4-4e9eb3cb3c23-kube-api-access-m7hgq\") pod \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.366978 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-config-data\") pod \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.366995 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-config-data\") pod \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367023 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-combined-ca-bundle\") pod \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367088 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-combined-ca-bundle\") pod \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367144 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-scripts\") pod \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367188 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-credential-keys\") pod \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\" (UID: \"4482a16d-c783-4609-9db4-4e9eb3cb3c23\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367234 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-logs\") pod \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367262 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhhqd\" (UniqueName: \"kubernetes.io/projected/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-kube-api-access-jhhqd\") pod \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.367333 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-scripts\") pod \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\" (UID: \"3d26f30b-efdc-40ef-9ae5-ca78e79aba41\") " Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.368413 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-logs" (OuterVolumeSpecName: "logs") pod "3d26f30b-efdc-40ef-9ae5-ca78e79aba41" (UID: "3d26f30b-efdc-40ef-9ae5-ca78e79aba41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.372689 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-scripts" (OuterVolumeSpecName: "scripts") pod "4482a16d-c783-4609-9db4-4e9eb3cb3c23" (UID: "4482a16d-c783-4609-9db4-4e9eb3cb3c23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.394271 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4482a16d-c783-4609-9db4-4e9eb3cb3c23" (UID: "4482a16d-c783-4609-9db4-4e9eb3cb3c23"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.394701 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-scripts" (OuterVolumeSpecName: "scripts") pod "3d26f30b-efdc-40ef-9ae5-ca78e79aba41" (UID: "3d26f30b-efdc-40ef-9ae5-ca78e79aba41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.398699 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4482a16d-c783-4609-9db4-4e9eb3cb3c23-kube-api-access-m7hgq" (OuterVolumeSpecName: "kube-api-access-m7hgq") pod "4482a16d-c783-4609-9db4-4e9eb3cb3c23" (UID: "4482a16d-c783-4609-9db4-4e9eb3cb3c23"). InnerVolumeSpecName "kube-api-access-m7hgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.399292 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-kube-api-access-jhhqd" (OuterVolumeSpecName: "kube-api-access-jhhqd") pod "3d26f30b-efdc-40ef-9ae5-ca78e79aba41" (UID: "3d26f30b-efdc-40ef-9ae5-ca78e79aba41"). InnerVolumeSpecName "kube-api-access-jhhqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.402406 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4482a16d-c783-4609-9db4-4e9eb3cb3c23" (UID: "4482a16d-c783-4609-9db4-4e9eb3cb3c23"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.403192 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gvbcw" event={"ID":"3d26f30b-efdc-40ef-9ae5-ca78e79aba41","Type":"ContainerDied","Data":"377e4e9171494829eb218364c92fa20a57c414340ee43107e92819698badf202"} Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.403231 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="377e4e9171494829eb218364c92fa20a57c414340ee43107e92819698badf202" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.403295 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gvbcw" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.406681 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d26f30b-efdc-40ef-9ae5-ca78e79aba41" (UID: "3d26f30b-efdc-40ef-9ae5-ca78e79aba41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.408479 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-config-data" (OuterVolumeSpecName: "config-data") pod "4482a16d-c783-4609-9db4-4e9eb3cb3c23" (UID: "4482a16d-c783-4609-9db4-4e9eb3cb3c23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.414122 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4482a16d-c783-4609-9db4-4e9eb3cb3c23" (UID: "4482a16d-c783-4609-9db4-4e9eb3cb3c23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.414847 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lgl42" event={"ID":"4482a16d-c783-4609-9db4-4e9eb3cb3c23","Type":"ContainerDied","Data":"cde776bb251d20603b7aa46b3a69b0cc3d92f86b92a1b97b6cc0a62bd0075ed3"} Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.414888 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cde776bb251d20603b7aa46b3a69b0cc3d92f86b92a1b97b6cc0a62bd0075ed3" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.414952 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lgl42" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.419917 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-config-data" (OuterVolumeSpecName: "config-data") pod "3d26f30b-efdc-40ef-9ae5-ca78e79aba41" (UID: "3d26f30b-efdc-40ef-9ae5-ca78e79aba41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469425 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhhqd\" (UniqueName: \"kubernetes.io/projected/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-kube-api-access-jhhqd\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469473 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469491 4817 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469509 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7hgq\" (UniqueName: \"kubernetes.io/projected/4482a16d-c783-4609-9db4-4e9eb3cb3c23-kube-api-access-m7hgq\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469526 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469543 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469559 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469575 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469592 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469608 4817 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4482a16d-c783-4609-9db4-4e9eb3cb3c23-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:21 crc kubenswrapper[4817]: I1128 14:47:21.469623 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d26f30b-efdc-40ef-9ae5-ca78e79aba41-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.409305 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6746568d76-zwnfd"] Nov 28 14:47:22 crc kubenswrapper[4817]: E1128 14:47:22.409655 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482a16d-c783-4609-9db4-4e9eb3cb3c23" containerName="keystone-bootstrap" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.409667 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482a16d-c783-4609-9db4-4e9eb3cb3c23" containerName="keystone-bootstrap" Nov 28 14:47:22 crc kubenswrapper[4817]: E1128 14:47:22.409675 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d26f30b-efdc-40ef-9ae5-ca78e79aba41" containerName="placement-db-sync" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.409680 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d26f30b-efdc-40ef-9ae5-ca78e79aba41" containerName="placement-db-sync" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.409875 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="4482a16d-c783-4609-9db4-4e9eb3cb3c23" containerName="keystone-bootstrap" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.409892 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d26f30b-efdc-40ef-9ae5-ca78e79aba41" containerName="placement-db-sync" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.410699 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.412864 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.413303 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.413648 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-kcxkk" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.414253 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.423076 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.426189 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmlbp" event={"ID":"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026","Type":"ContainerStarted","Data":"98b76e96e43924f6b4865883de4166204f1ee77ce977c706ea40861adb11191f"} Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.432012 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6746568d76-zwnfd"] Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494005 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-config-data\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494083 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sjvv\" (UniqueName: \"kubernetes.io/projected/4164fdcc-c963-4a08-9bd5-2c569d096180-kube-api-access-7sjvv\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494127 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-scripts\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494168 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4164fdcc-c963-4a08-9bd5-2c569d096180-logs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494188 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-internal-tls-certs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494270 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-combined-ca-bundle\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.494301 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-public-tls-certs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.524320 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c874c964f-4j6hm"] Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.530976 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.537597 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.538934 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.539061 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.539078 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.539236 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-knk6m" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.539270 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.559326 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c874c964f-4j6hm"] Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.561898 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wmlbp" podStartSLOduration=26.471867202 podStartE2EDuration="32.56188061s" podCreationTimestamp="2025-11-28 14:46:50 +0000 UTC" firstStartedPulling="2025-11-28 14:47:10.91541507 +0000 UTC m=+1053.503393336" lastFinishedPulling="2025-11-28 14:47:17.005428438 +0000 UTC m=+1059.593406744" observedRunningTime="2025-11-28 14:47:22.525412124 +0000 UTC m=+1065.113390390" watchObservedRunningTime="2025-11-28 14:47:22.56188061 +0000 UTC m=+1065.149858876" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596599 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sjvv\" (UniqueName: \"kubernetes.io/projected/4164fdcc-c963-4a08-9bd5-2c569d096180-kube-api-access-7sjvv\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596642 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-credential-keys\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596660 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-combined-ca-bundle\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596685 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-scripts\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596710 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-internal-tls-certs\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596741 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-public-tls-certs\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596761 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4164fdcc-c963-4a08-9bd5-2c569d096180-logs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596775 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-internal-tls-certs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596813 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-combined-ca-bundle\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596828 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-config-data\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596845 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-scripts\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596860 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-public-tls-certs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596906 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwpk4\" (UniqueName: \"kubernetes.io/projected/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-kube-api-access-vwpk4\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596938 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-fernet-keys\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.596956 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-config-data\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.600916 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4164fdcc-c963-4a08-9bd5-2c569d096180-logs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.605702 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-public-tls-certs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.606386 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-config-data\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.609427 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-internal-tls-certs\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.610262 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-combined-ca-bundle\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.615185 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4164fdcc-c963-4a08-9bd5-2c569d096180-scripts\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.646424 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sjvv\" (UniqueName: \"kubernetes.io/projected/4164fdcc-c963-4a08-9bd5-2c569d096180-kube-api-access-7sjvv\") pod \"placement-6746568d76-zwnfd\" (UID: \"4164fdcc-c963-4a08-9bd5-2c569d096180\") " pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.703585 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-config-data\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.703988 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-scripts\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.704055 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwpk4\" (UniqueName: \"kubernetes.io/projected/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-kube-api-access-vwpk4\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.704091 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-fernet-keys\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.704121 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-credential-keys\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.704139 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-combined-ca-bundle\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.704170 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-internal-tls-certs\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.704186 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-public-tls-certs\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.710586 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-config-data\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.713089 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-scripts\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.713475 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-public-tls-certs\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.715472 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-combined-ca-bundle\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.724172 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-credential-keys\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.724631 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-fernet-keys\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.725750 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-internal-tls-certs\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.730204 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.735337 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwpk4\" (UniqueName: \"kubernetes.io/projected/b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d-kube-api-access-vwpk4\") pod \"keystone-5c874c964f-4j6hm\" (UID: \"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d\") " pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:22 crc kubenswrapper[4817]: I1128 14:47:22.875264 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:23 crc kubenswrapper[4817]: I1128 14:47:23.174357 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6746568d76-zwnfd"] Nov 28 14:47:23 crc kubenswrapper[4817]: W1128 14:47:23.176519 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4164fdcc_c963_4a08_9bd5_2c569d096180.slice/crio-92775a4482c005709cc676b15495a8b760221aac97671ce66bf747c315945e05 WatchSource:0}: Error finding container 92775a4482c005709cc676b15495a8b760221aac97671ce66bf747c315945e05: Status 404 returned error can't find the container with id 92775a4482c005709cc676b15495a8b760221aac97671ce66bf747c315945e05 Nov 28 14:47:23 crc kubenswrapper[4817]: I1128 14:47:23.326498 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c874c964f-4j6hm"] Nov 28 14:47:23 crc kubenswrapper[4817]: W1128 14:47:23.342585 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3a6b7c5_6c6d_4ebc_bc5b_b9ab47e95d3d.slice/crio-cda0eea935ae34f92c7b00e8640948d85f82b06f8b9e712388f96bbbb65a85b4 WatchSource:0}: Error finding container cda0eea935ae34f92c7b00e8640948d85f82b06f8b9e712388f96bbbb65a85b4: Status 404 returned error can't find the container with id cda0eea935ae34f92c7b00e8640948d85f82b06f8b9e712388f96bbbb65a85b4 Nov 28 14:47:23 crc kubenswrapper[4817]: I1128 14:47:23.442203 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9f53839-25c4-4e8f-b9c7-1895ddff10d6","Type":"ContainerStarted","Data":"f4f22ca1b0a23418a0255001362927ff0e4b8f602cb4785a1133d170f894b648"} Nov 28 14:47:23 crc kubenswrapper[4817]: I1128 14:47:23.444495 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c874c964f-4j6hm" event={"ID":"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d","Type":"ContainerStarted","Data":"cda0eea935ae34f92c7b00e8640948d85f82b06f8b9e712388f96bbbb65a85b4"} Nov 28 14:47:23 crc kubenswrapper[4817]: I1128 14:47:23.447892 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6746568d76-zwnfd" event={"ID":"4164fdcc-c963-4a08-9bd5-2c569d096180","Type":"ContainerStarted","Data":"92775a4482c005709cc676b15495a8b760221aac97671ce66bf747c315945e05"} Nov 28 14:47:23 crc kubenswrapper[4817]: I1128 14:47:23.474147 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.474124844 podStartE2EDuration="11.474124844s" podCreationTimestamp="2025-11-28 14:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:23.465037158 +0000 UTC m=+1066.053015434" watchObservedRunningTime="2025-11-28 14:47:23.474124844 +0000 UTC m=+1066.062103110" Nov 28 14:47:24 crc kubenswrapper[4817]: E1128 14:47:24.036172 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6da990a2_1382_4707_8311_5dd46e4f537c.slice/crio-conmon-aa36b76a97ce87e35668be6b4c9bb5840bf2a30d6a66424da3c73491b117c116.scope\": RecentStats: unable to find data in memory cache]" Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.458114 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6746568d76-zwnfd" event={"ID":"4164fdcc-c963-4a08-9bd5-2c569d096180","Type":"ContainerStarted","Data":"3ec45de38c81f66136367fcd0039030c13671ab661856833b396179d09cdb96e"} Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.458454 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.458467 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6746568d76-zwnfd" event={"ID":"4164fdcc-c963-4a08-9bd5-2c569d096180","Type":"ContainerStarted","Data":"802fe9a068b249628a57bc086fa394b6a644c3a8d490fba1627af906722a310e"} Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.459994 4817 generic.go:334] "Generic (PLEG): container finished" podID="6da990a2-1382-4707-8311-5dd46e4f537c" containerID="aa36b76a97ce87e35668be6b4c9bb5840bf2a30d6a66424da3c73491b117c116" exitCode=0 Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.460038 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fbfcr" event={"ID":"6da990a2-1382-4707-8311-5dd46e4f537c","Type":"ContainerDied","Data":"aa36b76a97ce87e35668be6b4c9bb5840bf2a30d6a66424da3c73491b117c116"} Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.462154 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerStarted","Data":"cacb69c5846a15b08b4240e267546dc031e7dd959725b73c99f209add92bf579"} Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.465618 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mpzzx" event={"ID":"11e39ba1-49a5-4289-a913-10b87e82f4e0","Type":"ContainerStarted","Data":"91fb5eccf01242cd19f09770332a91907ca62745c86c2a4555cfbaea57beb469"} Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.468516 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c874c964f-4j6hm" event={"ID":"b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d","Type":"ContainerStarted","Data":"d22b010bddc6a8682dbc0df8d02543a26722d2d88819c02fcc72d4720a4eb3a7"} Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.468738 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.476567 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6746568d76-zwnfd" podStartSLOduration=2.47654455 podStartE2EDuration="2.47654455s" podCreationTimestamp="2025-11-28 14:47:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:24.476451617 +0000 UTC m=+1067.064429883" watchObservedRunningTime="2025-11-28 14:47:24.47654455 +0000 UTC m=+1067.064522816" Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.494652 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-mpzzx" podStartSLOduration=3.950367425 podStartE2EDuration="56.494632159s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="2025-11-28 14:46:31.3090017 +0000 UTC m=+1013.896979966" lastFinishedPulling="2025-11-28 14:47:23.853266434 +0000 UTC m=+1066.441244700" observedRunningTime="2025-11-28 14:47:24.491807979 +0000 UTC m=+1067.079786245" watchObservedRunningTime="2025-11-28 14:47:24.494632159 +0000 UTC m=+1067.082610425" Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.527437 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c874c964f-4j6hm" podStartSLOduration=2.527419173 podStartE2EDuration="2.527419173s" podCreationTimestamp="2025-11-28 14:47:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:24.511072627 +0000 UTC m=+1067.099050893" watchObservedRunningTime="2025-11-28 14:47:24.527419173 +0000 UTC m=+1067.115397439" Nov 28 14:47:24 crc kubenswrapper[4817]: I1128 14:47:24.650685 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gk926" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" probeResult="failure" output=< Nov 28 14:47:24 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 14:47:24 crc kubenswrapper[4817]: > Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.497903 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q2qk" event={"ID":"ca79eb0e-3393-4730-ba3a-c56fc3c6623f","Type":"ContainerStarted","Data":"4a042097560c269170cec7774bc7e066e6dfe0f4584a1b0e56eb100875839b04"} Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.511270 4817 generic.go:334] "Generic (PLEG): container finished" podID="5d2d10b9-2269-4654-a979-cf2c9a8c65dc" containerID="a9dce10085e4a2031130e9f98ffac108e9faf70fde37937df7241bc8a088d87d" exitCode=0 Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.511564 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fflrd" event={"ID":"5d2d10b9-2269-4654-a979-cf2c9a8c65dc","Type":"ContainerDied","Data":"a9dce10085e4a2031130e9f98ffac108e9faf70fde37937df7241bc8a088d87d"} Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.512535 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.526957 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-5q2qk" podStartSLOduration=5.453914791 podStartE2EDuration="57.526941377s" podCreationTimestamp="2025-11-28 14:46:28 +0000 UTC" firstStartedPulling="2025-11-28 14:46:31.307087932 +0000 UTC m=+1013.895066198" lastFinishedPulling="2025-11-28 14:47:23.380114518 +0000 UTC m=+1065.968092784" observedRunningTime="2025-11-28 14:47:25.521039581 +0000 UTC m=+1068.109017847" watchObservedRunningTime="2025-11-28 14:47:25.526941377 +0000 UTC m=+1068.114919643" Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.824554 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fbfcr" Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.965172 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-config-data\") pod \"6da990a2-1382-4707-8311-5dd46e4f537c\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.965528 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-combined-ca-bundle\") pod \"6da990a2-1382-4707-8311-5dd46e4f537c\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.965626 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw4g7\" (UniqueName: \"kubernetes.io/projected/6da990a2-1382-4707-8311-5dd46e4f537c-kube-api-access-dw4g7\") pod \"6da990a2-1382-4707-8311-5dd46e4f537c\" (UID: \"6da990a2-1382-4707-8311-5dd46e4f537c\") " Nov 28 14:47:25 crc kubenswrapper[4817]: I1128 14:47:25.988499 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da990a2-1382-4707-8311-5dd46e4f537c-kube-api-access-dw4g7" (OuterVolumeSpecName: "kube-api-access-dw4g7") pod "6da990a2-1382-4707-8311-5dd46e4f537c" (UID: "6da990a2-1382-4707-8311-5dd46e4f537c"). InnerVolumeSpecName "kube-api-access-dw4g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.003809 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6da990a2-1382-4707-8311-5dd46e4f537c" (UID: "6da990a2-1382-4707-8311-5dd46e4f537c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.043550 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-config-data" (OuterVolumeSpecName: "config-data") pod "6da990a2-1382-4707-8311-5dd46e4f537c" (UID: "6da990a2-1382-4707-8311-5dd46e4f537c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.068497 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.068539 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da990a2-1382-4707-8311-5dd46e4f537c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.068555 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw4g7\" (UniqueName: \"kubernetes.io/projected/6da990a2-1382-4707-8311-5dd46e4f537c-kube-api-access-dw4g7\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.523303 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-fbfcr" event={"ID":"6da990a2-1382-4707-8311-5dd46e4f537c","Type":"ContainerDied","Data":"87c5fbdb25d6896b406716797e4d96a1352c4fedf2c1c9d970c6af3330d7c2e5"} Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.525460 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87c5fbdb25d6896b406716797e4d96a1352c4fedf2c1c9d970c6af3330d7c2e5" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.523647 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-fbfcr" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.850320 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fflrd" Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.982323 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-combined-ca-bundle\") pod \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.982401 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-config\") pod \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.982540 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29sxk\" (UniqueName: \"kubernetes.io/projected/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-kube-api-access-29sxk\") pod \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\" (UID: \"5d2d10b9-2269-4654-a979-cf2c9a8c65dc\") " Nov 28 14:47:26 crc kubenswrapper[4817]: I1128 14:47:26.989024 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-kube-api-access-29sxk" (OuterVolumeSpecName: "kube-api-access-29sxk") pod "5d2d10b9-2269-4654-a979-cf2c9a8c65dc" (UID: "5d2d10b9-2269-4654-a979-cf2c9a8c65dc"). InnerVolumeSpecName "kube-api-access-29sxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.012838 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d2d10b9-2269-4654-a979-cf2c9a8c65dc" (UID: "5d2d10b9-2269-4654-a979-cf2c9a8c65dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.021429 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-config" (OuterVolumeSpecName: "config") pod "5d2d10b9-2269-4654-a979-cf2c9a8c65dc" (UID: "5d2d10b9-2269-4654-a979-cf2c9a8c65dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.084634 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.084668 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29sxk\" (UniqueName: \"kubernetes.io/projected/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-kube-api-access-29sxk\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.084678 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2d10b9-2269-4654-a979-cf2c9a8c65dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.547181 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fflrd" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.547695 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fflrd" event={"ID":"5d2d10b9-2269-4654-a979-cf2c9a8c65dc","Type":"ContainerDied","Data":"fdbb165a75232b4190996ebb3f5eb44a8214a7bc0c3422327de8683659382259"} Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.547744 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdbb165a75232b4190996ebb3f5eb44a8214a7bc0c3422327de8683659382259" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.549387 4817 generic.go:334] "Generic (PLEG): container finished" podID="11e39ba1-49a5-4289-a913-10b87e82f4e0" containerID="91fb5eccf01242cd19f09770332a91907ca62745c86c2a4555cfbaea57beb469" exitCode=0 Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.549418 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mpzzx" event={"ID":"11e39ba1-49a5-4289-a913-10b87e82f4e0","Type":"ContainerDied","Data":"91fb5eccf01242cd19f09770332a91907ca62745c86c2a4555cfbaea57beb469"} Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.707464 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hmc2c"] Nov 28 14:47:27 crc kubenswrapper[4817]: E1128 14:47:27.708346 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da990a2-1382-4707-8311-5dd46e4f537c" containerName="heat-db-sync" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.708372 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da990a2-1382-4707-8311-5dd46e4f537c" containerName="heat-db-sync" Nov 28 14:47:27 crc kubenswrapper[4817]: E1128 14:47:27.708404 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2d10b9-2269-4654-a979-cf2c9a8c65dc" containerName="neutron-db-sync" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.708414 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2d10b9-2269-4654-a979-cf2c9a8c65dc" containerName="neutron-db-sync" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.708654 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2d10b9-2269-4654-a979-cf2c9a8c65dc" containerName="neutron-db-sync" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.708677 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da990a2-1382-4707-8311-5dd46e4f537c" containerName="heat-db-sync" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.709977 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.719067 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hmc2c"] Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.803065 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.803108 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.803131 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.803154 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.803185 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-config\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.803583 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvmm7\" (UniqueName: \"kubernetes.io/projected/c016fabd-3bd6-4ef0-9687-01db2cb0db63-kube-api-access-lvmm7\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.861861 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c989bbc58-5p729"] Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.863512 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.867461 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.868056 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.868206 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.868317 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9hvz9" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.878732 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c989bbc58-5p729"] Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904760 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-combined-ca-bundle\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904827 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvmm7\" (UniqueName: \"kubernetes.io/projected/c016fabd-3bd6-4ef0-9687-01db2cb0db63-kube-api-access-lvmm7\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904860 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9hqj\" (UniqueName: \"kubernetes.io/projected/eb5a94ef-ba23-48a4-99e6-23f6764f8606-kube-api-access-f9hqj\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904916 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-config\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904956 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904975 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.904993 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-ovndb-tls-certs\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.905007 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.905022 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.905044 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-config\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.905078 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-httpd-config\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.907312 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.907832 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.908387 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-svc\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.908743 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.909042 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-config\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:27 crc kubenswrapper[4817]: I1128 14:47:27.927807 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvmm7\" (UniqueName: \"kubernetes.io/projected/c016fabd-3bd6-4ef0-9687-01db2cb0db63-kube-api-access-lvmm7\") pod \"dnsmasq-dns-55f844cf75-hmc2c\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.006518 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-httpd-config\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.006630 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-combined-ca-bundle\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.007650 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9hqj\" (UniqueName: \"kubernetes.io/projected/eb5a94ef-ba23-48a4-99e6-23f6764f8606-kube-api-access-f9hqj\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.007741 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-config\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.007833 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-ovndb-tls-certs\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.012452 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-ovndb-tls-certs\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.013507 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-combined-ca-bundle\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.016884 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-config\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.016934 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-httpd-config\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.029195 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9hqj\" (UniqueName: \"kubernetes.io/projected/eb5a94ef-ba23-48a4-99e6-23f6764f8606-kube-api-access-f9hqj\") pod \"neutron-c989bbc58-5p729\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.039035 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:28 crc kubenswrapper[4817]: I1128 14:47:28.181855 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.762691 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cc96c44df-rqchh"] Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.766127 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.768474 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.768578 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.778637 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cc96c44df-rqchh"] Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.845798 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdk9j\" (UniqueName: \"kubernetes.io/projected/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-kube-api-access-mdk9j\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.845938 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-internal-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.845987 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-config\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.846009 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-public-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.846041 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-combined-ca-bundle\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.846085 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-httpd-config\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.846167 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-ovndb-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948114 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-internal-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948188 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-config\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948212 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-public-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948246 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-combined-ca-bundle\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948287 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-httpd-config\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948338 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-ovndb-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.948398 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdk9j\" (UniqueName: \"kubernetes.io/projected/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-kube-api-access-mdk9j\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.960510 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-internal-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.960617 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-combined-ca-bundle\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.962701 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-public-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.970686 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-config\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.975465 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-ovndb-tls-certs\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.976205 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-httpd-config\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:29 crc kubenswrapper[4817]: I1128 14:47:29.984656 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdk9j\" (UniqueName: \"kubernetes.io/projected/0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9-kube-api-access-mdk9j\") pod \"neutron-6cc96c44df-rqchh\" (UID: \"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9\") " pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:30 crc kubenswrapper[4817]: I1128 14:47:30.088319 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:30 crc kubenswrapper[4817]: I1128 14:47:30.579992 4817 generic.go:334] "Generic (PLEG): container finished" podID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" containerID="4a042097560c269170cec7774bc7e066e6dfe0f4584a1b0e56eb100875839b04" exitCode=0 Nov 28 14:47:30 crc kubenswrapper[4817]: I1128 14:47:30.580047 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q2qk" event={"ID":"ca79eb0e-3393-4730-ba3a-c56fc3c6623f","Type":"ContainerDied","Data":"4a042097560c269170cec7774bc7e066e6dfe0f4584a1b0e56eb100875839b04"} Nov 28 14:47:30 crc kubenswrapper[4817]: I1128 14:47:30.933421 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:47:30 crc kubenswrapper[4817]: I1128 14:47:30.933815 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:47:30 crc kubenswrapper[4817]: I1128 14:47:30.999337 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:47:31 crc kubenswrapper[4817]: I1128 14:47:31.646737 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:47:31 crc kubenswrapper[4817]: I1128 14:47:31.707628 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wmlbp"] Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.422024 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.438035 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493380 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-config-data\") pod \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493497 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-db-sync-config-data\") pod \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493542 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-db-sync-config-data\") pod \"11e39ba1-49a5-4289-a913-10b87e82f4e0\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493559 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkg8l\" (UniqueName: \"kubernetes.io/projected/11e39ba1-49a5-4289-a913-10b87e82f4e0-kube-api-access-hkg8l\") pod \"11e39ba1-49a5-4289-a913-10b87e82f4e0\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493581 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-combined-ca-bundle\") pod \"11e39ba1-49a5-4289-a913-10b87e82f4e0\" (UID: \"11e39ba1-49a5-4289-a913-10b87e82f4e0\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493630 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-combined-ca-bundle\") pod \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493665 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-etc-machine-id\") pod \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493686 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfctv\" (UniqueName: \"kubernetes.io/projected/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-kube-api-access-pfctv\") pod \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493704 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-scripts\") pod \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\" (UID: \"ca79eb0e-3393-4730-ba3a-c56fc3c6623f\") " Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.493795 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ca79eb0e-3393-4730-ba3a-c56fc3c6623f" (UID: "ca79eb0e-3393-4730-ba3a-c56fc3c6623f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.494096 4817 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.499583 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-kube-api-access-pfctv" (OuterVolumeSpecName: "kube-api-access-pfctv") pod "ca79eb0e-3393-4730-ba3a-c56fc3c6623f" (UID: "ca79eb0e-3393-4730-ba3a-c56fc3c6623f"). InnerVolumeSpecName "kube-api-access-pfctv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.501307 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11e39ba1-49a5-4289-a913-10b87e82f4e0-kube-api-access-hkg8l" (OuterVolumeSpecName: "kube-api-access-hkg8l") pod "11e39ba1-49a5-4289-a913-10b87e82f4e0" (UID: "11e39ba1-49a5-4289-a913-10b87e82f4e0"). InnerVolumeSpecName "kube-api-access-hkg8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.501341 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ca79eb0e-3393-4730-ba3a-c56fc3c6623f" (UID: "ca79eb0e-3393-4730-ba3a-c56fc3c6623f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.501445 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "11e39ba1-49a5-4289-a913-10b87e82f4e0" (UID: "11e39ba1-49a5-4289-a913-10b87e82f4e0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.509069 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-scripts" (OuterVolumeSpecName: "scripts") pod "ca79eb0e-3393-4730-ba3a-c56fc3c6623f" (UID: "ca79eb0e-3393-4730-ba3a-c56fc3c6623f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.527147 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca79eb0e-3393-4730-ba3a-c56fc3c6623f" (UID: "ca79eb0e-3393-4730-ba3a-c56fc3c6623f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.536045 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11e39ba1-49a5-4289-a913-10b87e82f4e0" (UID: "11e39ba1-49a5-4289-a913-10b87e82f4e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.576736 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-config-data" (OuterVolumeSpecName: "config-data") pod "ca79eb0e-3393-4730-ba3a-c56fc3c6623f" (UID: "ca79eb0e-3393-4730-ba3a-c56fc3c6623f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597816 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597838 4817 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597849 4817 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597858 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkg8l\" (UniqueName: \"kubernetes.io/projected/11e39ba1-49a5-4289-a913-10b87e82f4e0-kube-api-access-hkg8l\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597867 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11e39ba1-49a5-4289-a913-10b87e82f4e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597875 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597989 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfctv\" (UniqueName: \"kubernetes.io/projected/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-kube-api-access-pfctv\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.597999 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca79eb0e-3393-4730-ba3a-c56fc3c6623f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.598405 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q2qk" event={"ID":"ca79eb0e-3393-4730-ba3a-c56fc3c6623f","Type":"ContainerDied","Data":"6090483594e822d0e62effb8ad101c82383a3ce6984418660e1ff7092d22fa9b"} Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.598436 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6090483594e822d0e62effb8ad101c82383a3ce6984418660e1ff7092d22fa9b" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.598484 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q2qk" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.600082 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mpzzx" event={"ID":"11e39ba1-49a5-4289-a913-10b87e82f4e0","Type":"ContainerDied","Data":"7bd0f1b7af817d105f67194cd022d46acf73e3a5f38a63930ad24b85e9d4a03a"} Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.600153 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bd0f1b7af817d105f67194cd022d46acf73e3a5f38a63930ad24b85e9d4a03a" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.600836 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mpzzx" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.693581 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.693617 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.726697 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.747232 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.904029 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:32 crc kubenswrapper[4817]: E1128 14:47:32.904526 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" containerName="cinder-db-sync" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.904550 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" containerName="cinder-db-sync" Nov 28 14:47:32 crc kubenswrapper[4817]: E1128 14:47:32.904584 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e39ba1-49a5-4289-a913-10b87e82f4e0" containerName="barbican-db-sync" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.904593 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e39ba1-49a5-4289-a913-10b87e82f4e0" containerName="barbican-db-sync" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.904865 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" containerName="cinder-db-sync" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.904901 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="11e39ba1-49a5-4289-a913-10b87e82f4e0" containerName="barbican-db-sync" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.906091 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.908520 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.908736 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.908902 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.911183 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dc9bg" Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.912848 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hmc2c"] Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.921811 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.979661 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c27cg"] Nov 28 14:47:32 crc kubenswrapper[4817]: I1128 14:47:32.981406 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011302 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-scripts\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011344 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011384 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011407 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-config\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011453 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82tt8\" (UniqueName: \"kubernetes.io/projected/cb687c4d-271e-4d0a-93f0-840148680b78-kube-api-access-82tt8\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011476 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011557 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c768eb61-2c73-41c4-a06c-88893fdb33e9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011597 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011637 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-svc\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011672 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z24k4\" (UniqueName: \"kubernetes.io/projected/c768eb61-2c73-41c4-a06c-88893fdb33e9-kube-api-access-z24k4\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011691 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.011752 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.014319 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c27cg"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113588 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-svc\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113635 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z24k4\" (UniqueName: \"kubernetes.io/projected/c768eb61-2c73-41c4-a06c-88893fdb33e9-kube-api-access-z24k4\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113657 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113699 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113728 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-scripts\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113745 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113763 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113782 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-config\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113807 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82tt8\" (UniqueName: \"kubernetes.io/projected/cb687c4d-271e-4d0a-93f0-840148680b78-kube-api-access-82tt8\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113824 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113880 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c768eb61-2c73-41c4-a06c-88893fdb33e9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.113905 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.119408 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c768eb61-2c73-41c4-a06c-88893fdb33e9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.119567 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.121669 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-config\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.122161 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.122709 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.125125 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-svc\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.127648 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.128078 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.129020 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-scripts\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.129337 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.147346 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82tt8\" (UniqueName: \"kubernetes.io/projected/cb687c4d-271e-4d0a-93f0-840148680b78-kube-api-access-82tt8\") pod \"dnsmasq-dns-b895b5785-c27cg\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.148514 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z24k4\" (UniqueName: \"kubernetes.io/projected/c768eb61-2c73-41c4-a06c-88893fdb33e9-kube-api-access-z24k4\") pod \"cinder-scheduler-0\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.188786 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.190803 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.195420 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.197150 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.218954 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adde3a48-0fe2-46e5-a82d-0ec41da371a9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.219005 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cldns\" (UniqueName: \"kubernetes.io/projected/adde3a48-0fe2-46e5-a82d-0ec41da371a9-kube-api-access-cldns\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.219034 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-scripts\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.219047 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adde3a48-0fe2-46e5-a82d-0ec41da371a9-logs\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.219071 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.219120 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data-custom\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.219134 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.233225 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.318937 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320573 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data-custom\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320618 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320692 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adde3a48-0fe2-46e5-a82d-0ec41da371a9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320740 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cldns\" (UniqueName: \"kubernetes.io/projected/adde3a48-0fe2-46e5-a82d-0ec41da371a9-kube-api-access-cldns\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320773 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adde3a48-0fe2-46e5-a82d-0ec41da371a9-logs\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320790 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-scripts\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.320818 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.323236 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adde3a48-0fe2-46e5-a82d-0ec41da371a9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.324202 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.324734 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adde3a48-0fe2-46e5-a82d-0ec41da371a9-logs\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.327147 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data-custom\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.327518 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-scripts\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.331539 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.344396 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cldns\" (UniqueName: \"kubernetes.io/projected/adde3a48-0fe2-46e5-a82d-0ec41da371a9-kube-api-access-cldns\") pod \"cinder-api-0\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.556901 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.585010 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-84d84dc879-8hskz"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.586588 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.590960 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.591365 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.591995 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-t9vqz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.626062 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-combined-ca-bundle\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.626104 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9022c280-17b2-48fc-b818-73797815a819-logs\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.626184 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-config-data\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.626242 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-config-data-custom\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.626270 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlcgz\" (UniqueName: \"kubernetes.io/projected/9022c280-17b2-48fc-b818-73797815a819-kube-api-access-rlcgz\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.627488 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wmlbp" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="registry-server" containerID="cri-o://98b76e96e43924f6b4865883de4166204f1ee77ce977c706ea40861adb11191f" gracePeriod=2 Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.628287 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.628485 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.637526 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84d84dc879-8hskz"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.657341 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7dd97866bb-qxt2v"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.667233 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.671935 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.690782 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7dd97866bb-qxt2v"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.706599 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c27cg"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.728817 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-config-data-custom\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.728858 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlcgz\" (UniqueName: \"kubernetes.io/projected/9022c280-17b2-48fc-b818-73797815a819-kube-api-access-rlcgz\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.728952 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cc501c6-50dd-4f39-b19b-20f10e6b9997-logs\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.728976 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-combined-ca-bundle\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.728996 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9022c280-17b2-48fc-b818-73797815a819-logs\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.729013 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-combined-ca-bundle\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.729033 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48fkl\" (UniqueName: \"kubernetes.io/projected/7cc501c6-50dd-4f39-b19b-20f10e6b9997-kube-api-access-48fkl\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.729073 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-config-data\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.729109 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-config-data\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.729146 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-config-data-custom\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.734895 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9022c280-17b2-48fc-b818-73797815a819-logs\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.739959 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-combined-ca-bundle\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.744348 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-config-data-custom\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.756955 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9022c280-17b2-48fc-b818-73797815a819-config-data\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.770399 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-8cwtv"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.772583 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.775248 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlcgz\" (UniqueName: \"kubernetes.io/projected/9022c280-17b2-48fc-b818-73797815a819-kube-api-access-rlcgz\") pod \"barbican-worker-84d84dc879-8hskz\" (UID: \"9022c280-17b2-48fc-b818-73797815a819\") " pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.782158 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-8cwtv"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.830854 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cc501c6-50dd-4f39-b19b-20f10e6b9997-logs\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.830914 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-combined-ca-bundle\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.830936 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48fkl\" (UniqueName: \"kubernetes.io/projected/7cc501c6-50dd-4f39-b19b-20f10e6b9997-kube-api-access-48fkl\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.830973 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-config-data\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.833090 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cc501c6-50dd-4f39-b19b-20f10e6b9997-logs\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.834265 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-config-data-custom\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.843927 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-combined-ca-bundle\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.845262 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-config-data-custom\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.846201 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cc501c6-50dd-4f39-b19b-20f10e6b9997-config-data\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.860054 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48fkl\" (UniqueName: \"kubernetes.io/projected/7cc501c6-50dd-4f39-b19b-20f10e6b9997-kube-api-access-48fkl\") pod \"barbican-keystone-listener-7dd97866bb-qxt2v\" (UID: \"7cc501c6-50dd-4f39-b19b-20f10e6b9997\") " pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.889351 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7b4d5d6b8b-8n4xc"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.891149 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.894393 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.897342 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b4d5d6b8b-8n4xc"] Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937505 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84d84dc879-8hskz" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937808 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqktf\" (UniqueName: \"kubernetes.io/projected/137e13fa-8f5d-4076-8de7-57c8284651dd-kube-api-access-kqktf\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937865 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937898 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937927 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9wv7\" (UniqueName: \"kubernetes.io/projected/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-kube-api-access-p9wv7\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937962 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data-custom\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.937982 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-combined-ca-bundle\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.938152 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.938183 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-config\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.938204 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137e13fa-8f5d-4076-8de7-57c8284651dd-logs\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.938294 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:33 crc kubenswrapper[4817]: I1128 14:47:33.938348 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.039641 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data-custom\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040428 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-combined-ca-bundle\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040484 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040506 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-config\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040524 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137e13fa-8f5d-4076-8de7-57c8284651dd-logs\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040579 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040613 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040658 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqktf\" (UniqueName: \"kubernetes.io/projected/137e13fa-8f5d-4076-8de7-57c8284651dd-kube-api-access-kqktf\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040681 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040702 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.040742 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9wv7\" (UniqueName: \"kubernetes.io/projected/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-kube-api-access-p9wv7\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.041536 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-config\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.041785 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.041858 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137e13fa-8f5d-4076-8de7-57c8284651dd-logs\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.041979 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.042348 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.042538 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.056386 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-combined-ca-bundle\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.067025 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data-custom\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.070879 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.105277 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqktf\" (UniqueName: \"kubernetes.io/projected/137e13fa-8f5d-4076-8de7-57c8284651dd-kube-api-access-kqktf\") pod \"barbican-api-7b4d5d6b8b-8n4xc\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.116108 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.126162 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9wv7\" (UniqueName: \"kubernetes.io/projected/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-kube-api-access-p9wv7\") pod \"dnsmasq-dns-5c9776ccc5-8cwtv\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.193571 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.230534 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:34 crc kubenswrapper[4817]: E1128 14:47:34.329491 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.657832 4817 generic.go:334] "Generic (PLEG): container finished" podID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerID="98b76e96e43924f6b4865883de4166204f1ee77ce977c706ea40861adb11191f" exitCode=0 Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.658448 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmlbp" event={"ID":"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026","Type":"ContainerDied","Data":"98b76e96e43924f6b4865883de4166204f1ee77ce977c706ea40861adb11191f"} Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.658472 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wmlbp" event={"ID":"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026","Type":"ContainerDied","Data":"429ad04dd947fe6d2dbeebaa9bbf17b4d031d828ae7c6d00d076b3abc0c32659"} Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.658482 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="429ad04dd947fe6d2dbeebaa9bbf17b4d031d828ae7c6d00d076b3abc0c32659" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.661836 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="ceilometer-notification-agent" containerID="cri-o://9c043a562b41a41ac20a4fc97fc6c7620409f0425c95c6f82e3900204828bea3" gracePeriod=30 Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.661967 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerStarted","Data":"cc5ce7039dec187f7ab155832429a56421d01dca1b22c0f61f8d420de16056fe"} Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.662039 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.662167 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="proxy-httpd" containerID="cri-o://cc5ce7039dec187f7ab155832429a56421d01dca1b22c0f61f8d420de16056fe" gracePeriod=30 Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.662224 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="sg-core" containerID="cri-o://cacb69c5846a15b08b4240e267546dc031e7dd959725b73c99f209add92bf579" gracePeriod=30 Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.686480 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gk926" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" probeResult="failure" output=< Nov 28 14:47:34 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 14:47:34 crc kubenswrapper[4817]: > Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.718090 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.862087 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cc96c44df-rqchh"] Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.873532 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-catalog-content\") pod \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.873617 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-utilities\") pod \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.873678 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmfqf\" (UniqueName: \"kubernetes.io/projected/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-kube-api-access-nmfqf\") pod \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\" (UID: \"c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026\") " Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.880341 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-kube-api-access-nmfqf" (OuterVolumeSpecName: "kube-api-access-nmfqf") pod "c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" (UID: "c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026"). InnerVolumeSpecName "kube-api-access-nmfqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.881342 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-utilities" (OuterVolumeSpecName: "utilities") pod "c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" (UID: "c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.937531 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" (UID: "c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.977783 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.977817 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:34 crc kubenswrapper[4817]: I1128 14:47:34.977831 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmfqf\" (UniqueName: \"kubernetes.io/projected/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026-kube-api-access-nmfqf\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.142857 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c27cg"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.168573 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.177983 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.188388 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hmc2c"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.213317 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84d84dc879-8hskz"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.242365 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.369966 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b4d5d6b8b-8n4xc"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.448778 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7dd97866bb-qxt2v"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.474502 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-8cwtv"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.798878 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adde3a48-0fe2-46e5-a82d-0ec41da371a9","Type":"ContainerStarted","Data":"fc5d470725d9f6a0f985db112171be8871574d16249ca36b04a94b8e6d87d8a2"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.800294 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc96c44df-rqchh" event={"ID":"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9","Type":"ContainerStarted","Data":"244c2641f5d8b45882b2a72ac0d238c4fefeb2c2c8637b874299ef41ccc279ee"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.800327 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc96c44df-rqchh" event={"ID":"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9","Type":"ContainerStarted","Data":"a0b33f80a6e05d95ee3b4c3d74f063490a10abd89b76713d3cd844f130e7ea31"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.807806 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" event={"ID":"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42","Type":"ContainerStarted","Data":"dba1f683c2eab8e31f755cc2ac2cead7f2e085a553b99ea9d8ff10ebaa20f9f2"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.814394 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" event={"ID":"137e13fa-8f5d-4076-8de7-57c8284651dd","Type":"ContainerStarted","Data":"b1597ad11344b60edf23ae90f43ff30b7a52c77c7c02d2703ebb5cb5e8fecfb3"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.827033 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" event={"ID":"7cc501c6-50dd-4f39-b19b-20f10e6b9997","Type":"ContainerStarted","Data":"5b7809b2b172de88af42224b151bc4d0b86c858108aae31ca55ed0999baffad4"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.831496 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c768eb61-2c73-41c4-a06c-88893fdb33e9","Type":"ContainerStarted","Data":"2a21fd6037762090157542f448c60103103d462668ba4d10d9f735063a82a9fc"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.835701 4817 generic.go:334] "Generic (PLEG): container finished" podID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerID="cc5ce7039dec187f7ab155832429a56421d01dca1b22c0f61f8d420de16056fe" exitCode=0 Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.835745 4817 generic.go:334] "Generic (PLEG): container finished" podID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerID="cacb69c5846a15b08b4240e267546dc031e7dd959725b73c99f209add92bf579" exitCode=2 Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.835783 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerDied","Data":"cc5ce7039dec187f7ab155832429a56421d01dca1b22c0f61f8d420de16056fe"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.835809 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerDied","Data":"cacb69c5846a15b08b4240e267546dc031e7dd959725b73c99f209add92bf579"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.836781 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-c27cg" event={"ID":"cb687c4d-271e-4d0a-93f0-840148680b78","Type":"ContainerStarted","Data":"e49adf9d1c541bf975d56a9f873ebb297d67646d07cb88ee90f5a0001c146bc5"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.840214 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84d84dc879-8hskz" event={"ID":"9022c280-17b2-48fc-b818-73797815a819","Type":"ContainerStarted","Data":"7f9388cd9a46cf8e9a2a392468e57115c4362085b8232e0a01ceb6fd03905d57"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.846484 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.846517 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.847454 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" event={"ID":"c016fabd-3bd6-4ef0-9687-01db2cb0db63","Type":"ContainerStarted","Data":"d2740bd2a1dabe4198836617548eb151c4bbc39b6e190754d6cafce0f13cf33f"} Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.847535 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wmlbp" Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.965704 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c989bbc58-5p729"] Nov 28 14:47:35 crc kubenswrapper[4817]: I1128 14:47:35.990315 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wmlbp"] Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.001106 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wmlbp"] Nov 28 14:47:36 crc kubenswrapper[4817]: W1128 14:47:36.008774 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb5a94ef_ba23_48a4_99e6_23f6764f8606.slice/crio-cd0eea617e090630850a3e62e6a41d2477bc998c1709306bc10579df914cf233 WatchSource:0}: Error finding container cd0eea617e090630850a3e62e6a41d2477bc998c1709306bc10579df914cf233: Status 404 returned error can't find the container with id cd0eea617e090630850a3e62e6a41d2477bc998c1709306bc10579df914cf233 Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.617890 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.638979 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.742209 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-config\") pod \"cb687c4d-271e-4d0a-93f0-840148680b78\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.742281 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-nb\") pod \"cb687c4d-271e-4d0a-93f0-840148680b78\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.742332 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-svc\") pod \"cb687c4d-271e-4d0a-93f0-840148680b78\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.742407 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82tt8\" (UniqueName: \"kubernetes.io/projected/cb687c4d-271e-4d0a-93f0-840148680b78-kube-api-access-82tt8\") pod \"cb687c4d-271e-4d0a-93f0-840148680b78\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.742444 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-sb\") pod \"cb687c4d-271e-4d0a-93f0-840148680b78\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.742459 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-swift-storage-0\") pod \"cb687c4d-271e-4d0a-93f0-840148680b78\" (UID: \"cb687c4d-271e-4d0a-93f0-840148680b78\") " Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.753392 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb687c4d-271e-4d0a-93f0-840148680b78-kube-api-access-82tt8" (OuterVolumeSpecName: "kube-api-access-82tt8") pod "cb687c4d-271e-4d0a-93f0-840148680b78" (UID: "cb687c4d-271e-4d0a-93f0-840148680b78"). InnerVolumeSpecName "kube-api-access-82tt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.772154 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cb687c4d-271e-4d0a-93f0-840148680b78" (UID: "cb687c4d-271e-4d0a-93f0-840148680b78"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.787874 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-config" (OuterVolumeSpecName: "config") pod "cb687c4d-271e-4d0a-93f0-840148680b78" (UID: "cb687c4d-271e-4d0a-93f0-840148680b78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.799255 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cb687c4d-271e-4d0a-93f0-840148680b78" (UID: "cb687c4d-271e-4d0a-93f0-840148680b78"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.804402 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cb687c4d-271e-4d0a-93f0-840148680b78" (UID: "cb687c4d-271e-4d0a-93f0-840148680b78"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.805184 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cb687c4d-271e-4d0a-93f0-840148680b78" (UID: "cb687c4d-271e-4d0a-93f0-840148680b78"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.845285 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.845316 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.845327 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.845336 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82tt8\" (UniqueName: \"kubernetes.io/projected/cb687c4d-271e-4d0a-93f0-840148680b78-kube-api-access-82tt8\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.845345 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.845353 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb687c4d-271e-4d0a-93f0-840148680b78-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.884565 4817 generic.go:334] "Generic (PLEG): container finished" podID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerID="2d8b0fd8a128f6adbee545e6ae5a4facd16d48d151450327e0dc6ea471e99ab2" exitCode=0 Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.884711 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" event={"ID":"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42","Type":"ContainerDied","Data":"2d8b0fd8a128f6adbee545e6ae5a4facd16d48d151450327e0dc6ea471e99ab2"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.900737 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" event={"ID":"137e13fa-8f5d-4076-8de7-57c8284651dd","Type":"ContainerStarted","Data":"ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.900780 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" event={"ID":"137e13fa-8f5d-4076-8de7-57c8284651dd","Type":"ContainerStarted","Data":"0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.902305 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.902353 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.911452 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adde3a48-0fe2-46e5-a82d-0ec41da371a9","Type":"ContainerStarted","Data":"b93b7e5375b5813689ae7b329737e52a9e07bf7b0cc619d366018dd0a4745064"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.914286 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc96c44df-rqchh" event={"ID":"0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9","Type":"ContainerStarted","Data":"7cfb00f720fdd7dd584304c0de122670aff4edb54e41edf704ecb02f842711c6"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.914435 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.916596 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c989bbc58-5p729" event={"ID":"eb5a94ef-ba23-48a4-99e6-23f6764f8606","Type":"ContainerStarted","Data":"daee428c9011ec401bff6caa7e5087bc7c3a041e8f72ad71a11316503a44d3fe"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.916635 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c989bbc58-5p729" event={"ID":"eb5a94ef-ba23-48a4-99e6-23f6764f8606","Type":"ContainerStarted","Data":"cd0eea617e090630850a3e62e6a41d2477bc998c1709306bc10579df914cf233"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.919046 4817 generic.go:334] "Generic (PLEG): container finished" podID="cb687c4d-271e-4d0a-93f0-840148680b78" containerID="ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2" exitCode=0 Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.919084 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-c27cg" event={"ID":"cb687c4d-271e-4d0a-93f0-840148680b78","Type":"ContainerDied","Data":"e49adf9d1c541bf975d56a9f873ebb297d67646d07cb88ee90f5a0001c146bc5"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.919099 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-c27cg" event={"ID":"cb687c4d-271e-4d0a-93f0-840148680b78","Type":"ContainerDied","Data":"ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.919114 4817 scope.go:117] "RemoveContainer" containerID="ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.919209 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c27cg" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.922983 4817 generic.go:334] "Generic (PLEG): container finished" podID="c016fabd-3bd6-4ef0-9687-01db2cb0db63" containerID="9fcfdbdb8e7036edd1f3551ce596091476cc61a0eadb14ebb67190fb6ce50b95" exitCode=0 Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.923072 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.923157 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" event={"ID":"c016fabd-3bd6-4ef0-9687-01db2cb0db63","Type":"ContainerDied","Data":"9fcfdbdb8e7036edd1f3551ce596091476cc61a0eadb14ebb67190fb6ce50b95"} Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.944395 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" podStartSLOduration=3.944375466 podStartE2EDuration="3.944375466s" podCreationTimestamp="2025-11-28 14:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:36.93126589 +0000 UTC m=+1079.519244156" watchObservedRunningTime="2025-11-28 14:47:36.944375466 +0000 UTC m=+1079.532353752" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.957365 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 14:47:36 crc kubenswrapper[4817]: I1128 14:47:36.978577 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cc96c44df-rqchh" podStartSLOduration=7.978561415 podStartE2EDuration="7.978561415s" podCreationTimestamp="2025-11-28 14:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:36.976471574 +0000 UTC m=+1079.564449840" watchObservedRunningTime="2025-11-28 14:47:36.978561415 +0000 UTC m=+1079.566539681" Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.031870 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c27cg"] Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.045741 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c27cg"] Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.815048 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" path="/var/lib/kubelet/pods/c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026/volumes" Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.816137 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb687c4d-271e-4d0a-93f0-840148680b78" path="/var/lib/kubelet/pods/cb687c4d-271e-4d0a-93f0-840148680b78/volumes" Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.826081 4817 scope.go:117] "RemoveContainer" containerID="ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2" Nov 28 14:47:37 crc kubenswrapper[4817]: E1128 14:47:37.826568 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2\": container with ID starting with ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2 not found: ID does not exist" containerID="ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2" Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.826601 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2"} err="failed to get container status \"ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2\": rpc error: code = NotFound desc = could not find container \"ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2\": container with ID starting with ee8eed5a09066381a66d0b5ed7a58d406c771866a6ae9fcdf26a2745a13b73a2 not found: ID does not exist" Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.953158 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api-log" containerID="cri-o://b93b7e5375b5813689ae7b329737e52a9e07bf7b0cc619d366018dd0a4745064" gracePeriod=30 Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.953503 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adde3a48-0fe2-46e5-a82d-0ec41da371a9","Type":"ContainerStarted","Data":"ce80e22ea9539afb7cc7dd7cd0b4aeea68fca33842a72a3c73a1262ba946fd63"} Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.953541 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.953852 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api" containerID="cri-o://ce80e22ea9539afb7cc7dd7cd0b4aeea68fca33842a72a3c73a1262ba946fd63" gracePeriod=30 Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.958130 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c768eb61-2c73-41c4-a06c-88893fdb33e9","Type":"ContainerStarted","Data":"bb6302042834a4d5c2a8a286e7af72073444ddb3a2ac133a8c603462430626a0"} Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.974878 4817 generic.go:334] "Generic (PLEG): container finished" podID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerID="9c043a562b41a41ac20a4fc97fc6c7620409f0425c95c6f82e3900204828bea3" exitCode=0 Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.974948 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerDied","Data":"9c043a562b41a41ac20a4fc97fc6c7620409f0425c95c6f82e3900204828bea3"} Nov 28 14:47:37 crc kubenswrapper[4817]: I1128 14:47:37.984906 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.984890847 podStartE2EDuration="4.984890847s" podCreationTimestamp="2025-11-28 14:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:37.975322969 +0000 UTC m=+1080.563301235" watchObservedRunningTime="2025-11-28 14:47:37.984890847 +0000 UTC m=+1080.572869113" Nov 28 14:47:38 crc kubenswrapper[4817]: I1128 14:47:38.910993 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.019991 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-svc\") pod \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.020045 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-swift-storage-0\") pod \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.020100 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-config\") pod \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.020180 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-nb\") pod \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.020204 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-sb\") pod \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.020327 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvmm7\" (UniqueName: \"kubernetes.io/projected/c016fabd-3bd6-4ef0-9687-01db2cb0db63-kube-api-access-lvmm7\") pod \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\" (UID: \"c016fabd-3bd6-4ef0-9687-01db2cb0db63\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.025529 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c016fabd-3bd6-4ef0-9687-01db2cb0db63-kube-api-access-lvmm7" (OuterVolumeSpecName: "kube-api-access-lvmm7") pod "c016fabd-3bd6-4ef0-9687-01db2cb0db63" (UID: "c016fabd-3bd6-4ef0-9687-01db2cb0db63"). InnerVolumeSpecName "kube-api-access-lvmm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.025659 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"49a012fe-0369-4c46-89e8-0c0062f46c33","Type":"ContainerDied","Data":"60c7413466656756c933ce65e921123036d121b753c2c3d9be6716cc778b7ce9"} Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.025696 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60c7413466656756c933ce65e921123036d121b753c2c3d9be6716cc778b7ce9" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.038282 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c989bbc58-5p729" event={"ID":"eb5a94ef-ba23-48a4-99e6-23f6764f8606","Type":"ContainerStarted","Data":"502f3a32af05550c174a7e7b618b7b84a04508d0b5bc48c2e2d5d965bbe03808"} Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.039467 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.052157 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c016fabd-3bd6-4ef0-9687-01db2cb0db63" (UID: "c016fabd-3bd6-4ef0-9687-01db2cb0db63"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.056456 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" event={"ID":"c016fabd-3bd6-4ef0-9687-01db2cb0db63","Type":"ContainerDied","Data":"d2740bd2a1dabe4198836617548eb151c4bbc39b6e190754d6cafce0f13cf33f"} Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.056517 4817 scope.go:117] "RemoveContainer" containerID="9fcfdbdb8e7036edd1f3551ce596091476cc61a0eadb14ebb67190fb6ce50b95" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.056833 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-hmc2c" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.062843 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c016fabd-3bd6-4ef0-9687-01db2cb0db63" (UID: "c016fabd-3bd6-4ef0-9687-01db2cb0db63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.072488 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.074694 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-config" (OuterVolumeSpecName: "config") pod "c016fabd-3bd6-4ef0-9687-01db2cb0db63" (UID: "c016fabd-3bd6-4ef0-9687-01db2cb0db63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.077508 4817 generic.go:334] "Generic (PLEG): container finished" podID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerID="ce80e22ea9539afb7cc7dd7cd0b4aeea68fca33842a72a3c73a1262ba946fd63" exitCode=0 Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.077540 4817 generic.go:334] "Generic (PLEG): container finished" podID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerID="b93b7e5375b5813689ae7b329737e52a9e07bf7b0cc619d366018dd0a4745064" exitCode=143 Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.078591 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adde3a48-0fe2-46e5-a82d-0ec41da371a9","Type":"ContainerDied","Data":"ce80e22ea9539afb7cc7dd7cd0b4aeea68fca33842a72a3c73a1262ba946fd63"} Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.078640 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adde3a48-0fe2-46e5-a82d-0ec41da371a9","Type":"ContainerDied","Data":"b93b7e5375b5813689ae7b329737e52a9e07bf7b0cc619d366018dd0a4745064"} Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.105565 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c989bbc58-5p729" podStartSLOduration=12.10554582 podStartE2EDuration="12.10554582s" podCreationTimestamp="2025-11-28 14:47:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:39.063769792 +0000 UTC m=+1081.651748078" watchObservedRunningTime="2025-11-28 14:47:39.10554582 +0000 UTC m=+1081.693524086" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.124797 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvmm7\" (UniqueName: \"kubernetes.io/projected/c016fabd-3bd6-4ef0-9687-01db2cb0db63-kube-api-access-lvmm7\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.124851 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.124864 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.124875 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.177769 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c016fabd-3bd6-4ef0-9687-01db2cb0db63" (UID: "c016fabd-3bd6-4ef0-9687-01db2cb0db63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.181785 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c016fabd-3bd6-4ef0-9687-01db2cb0db63" (UID: "c016fabd-3bd6-4ef0-9687-01db2cb0db63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226177 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-combined-ca-bundle\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226310 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-sg-core-conf-yaml\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226334 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-config-data\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226430 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-scripts\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226465 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx8qd\" (UniqueName: \"kubernetes.io/projected/49a012fe-0369-4c46-89e8-0c0062f46c33-kube-api-access-gx8qd\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226489 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-log-httpd\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226519 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-run-httpd\") pod \"49a012fe-0369-4c46-89e8-0c0062f46c33\" (UID: \"49a012fe-0369-4c46-89e8-0c0062f46c33\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226958 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.226969 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c016fabd-3bd6-4ef0-9687-01db2cb0db63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.227610 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.227889 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.244531 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-scripts" (OuterVolumeSpecName: "scripts") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.248047 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a012fe-0369-4c46-89e8-0c0062f46c33-kube-api-access-gx8qd" (OuterVolumeSpecName: "kube-api-access-gx8qd") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "kube-api-access-gx8qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.270563 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.279747 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.331005 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.331055 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx8qd\" (UniqueName: \"kubernetes.io/projected/49a012fe-0369-4c46-89e8-0c0062f46c33-kube-api-access-gx8qd\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.331087 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.331098 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/49a012fe-0369-4c46-89e8-0c0062f46c33-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.331109 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.340665 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.428798 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7fb9577b84-z8dsl"] Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429231 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c016fabd-3bd6-4ef0-9687-01db2cb0db63" containerName="init" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429244 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c016fabd-3bd6-4ef0-9687-01db2cb0db63" containerName="init" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429257 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="extract-utilities" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429264 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="extract-utilities" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429281 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="proxy-httpd" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429289 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="proxy-httpd" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429309 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="extract-content" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429315 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="extract-content" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429330 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="registry-server" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429337 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="registry-server" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429346 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api-log" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429354 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api-log" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429380 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429388 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429403 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb687c4d-271e-4d0a-93f0-840148680b78" containerName="init" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429410 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb687c4d-271e-4d0a-93f0-840148680b78" containerName="init" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429424 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="ceilometer-notification-agent" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429433 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="ceilometer-notification-agent" Nov 28 14:47:39 crc kubenswrapper[4817]: E1128 14:47:39.429452 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="sg-core" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429459 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="sg-core" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429614 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb687c4d-271e-4d0a-93f0-840148680b78" containerName="init" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429628 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c016fabd-3bd6-4ef0-9687-01db2cb0db63" containerName="init" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429639 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="ceilometer-notification-agent" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429652 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="proxy-httpd" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429659 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a2eeae-bde9-4fc9-9fc6-7cf765c8e026" containerName="registry-server" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429669 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" containerName="sg-core" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429675 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.429683 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" containerName="cinder-api-log" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.431244 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.432621 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.432692 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cldns\" (UniqueName: \"kubernetes.io/projected/adde3a48-0fe2-46e5-a82d-0ec41da371a9-kube-api-access-cldns\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.432789 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adde3a48-0fe2-46e5-a82d-0ec41da371a9-etc-machine-id\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.432887 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-combined-ca-bundle\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.432939 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data-custom\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.432979 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adde3a48-0fe2-46e5-a82d-0ec41da371a9-logs\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.433000 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-scripts\") pod \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\" (UID: \"adde3a48-0fe2-46e5-a82d-0ec41da371a9\") " Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.433356 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.434328 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adde3a48-0fe2-46e5-a82d-0ec41da371a9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.435931 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-config-data" (OuterVolumeSpecName: "config-data") pod "49a012fe-0369-4c46-89e8-0c0062f46c33" (UID: "49a012fe-0369-4c46-89e8-0c0062f46c33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.437438 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.437736 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.437800 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.438148 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adde3a48-0fe2-46e5-a82d-0ec41da371a9-logs" (OuterVolumeSpecName: "logs") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.447532 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adde3a48-0fe2-46e5-a82d-0ec41da371a9-kube-api-access-cldns" (OuterVolumeSpecName: "kube-api-access-cldns") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "kube-api-access-cldns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.454897 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-scripts" (OuterVolumeSpecName: "scripts") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.482374 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fb9577b84-z8dsl"] Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.503550 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.515808 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hmc2c"] Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.524779 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-hmc2c"] Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.528565 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data" (OuterVolumeSpecName: "config-data") pod "adde3a48-0fe2-46e5-a82d-0ec41da371a9" (UID: "adde3a48-0fe2-46e5-a82d-0ec41da371a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537553 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15698a16-9f4e-4552-b729-27969466063d-logs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537655 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-config-data-custom\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537754 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-public-tls-certs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537785 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-combined-ca-bundle\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537812 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-config-data\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537844 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmtdr\" (UniqueName: \"kubernetes.io/projected/15698a16-9f4e-4552-b729-27969466063d-kube-api-access-gmtdr\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537870 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-internal-tls-certs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537933 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537953 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a012fe-0369-4c46-89e8-0c0062f46c33-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537965 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537976 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adde3a48-0fe2-46e5-a82d-0ec41da371a9-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537987 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.537998 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adde3a48-0fe2-46e5-a82d-0ec41da371a9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.538011 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cldns\" (UniqueName: \"kubernetes.io/projected/adde3a48-0fe2-46e5-a82d-0ec41da371a9-kube-api-access-cldns\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.538024 4817 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adde3a48-0fe2-46e5-a82d-0ec41da371a9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639146 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15698a16-9f4e-4552-b729-27969466063d-logs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639545 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-config-data-custom\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639602 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-public-tls-certs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639626 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-combined-ca-bundle\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639646 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-config-data\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639670 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmtdr\" (UniqueName: \"kubernetes.io/projected/15698a16-9f4e-4552-b729-27969466063d-kube-api-access-gmtdr\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.639693 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-internal-tls-certs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.640223 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15698a16-9f4e-4552-b729-27969466063d-logs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.646453 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-combined-ca-bundle\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.647608 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-public-tls-certs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.647898 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-config-data-custom\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.650052 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-config-data\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.655280 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15698a16-9f4e-4552-b729-27969466063d-internal-tls-certs\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.659585 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmtdr\" (UniqueName: \"kubernetes.io/projected/15698a16-9f4e-4552-b729-27969466063d-kube-api-access-gmtdr\") pod \"barbican-api-7fb9577b84-z8dsl\" (UID: \"15698a16-9f4e-4552-b729-27969466063d\") " pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.747750 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c016fabd-3bd6-4ef0-9687-01db2cb0db63" path="/var/lib/kubelet/pods/c016fabd-3bd6-4ef0-9687-01db2cb0db63/volumes" Nov 28 14:47:39 crc kubenswrapper[4817]: I1128 14:47:39.777924 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.089230 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84d84dc879-8hskz" event={"ID":"9022c280-17b2-48fc-b818-73797815a819","Type":"ContainerStarted","Data":"4d28490af5c79564da97574dd588e22872255f7ad1890014af4ec501a49c57a1"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.089460 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84d84dc879-8hskz" event={"ID":"9022c280-17b2-48fc-b818-73797815a819","Type":"ContainerStarted","Data":"a91d3dee0c5170b2418bfbd271d5328955e24a146f93257ede26b3a5ad6484e8"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.096034 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" event={"ID":"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42","Type":"ContainerStarted","Data":"12296b6bbd7a948c76c48f76ea624ecf6b21180469d50c877b35fa3771591782"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.097098 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.098795 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" event={"ID":"7cc501c6-50dd-4f39-b19b-20f10e6b9997","Type":"ContainerStarted","Data":"bd1ac10e636ca461697b799724c56733b8db58e0b355ba953b97972e428d4c50"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.098828 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" event={"ID":"7cc501c6-50dd-4f39-b19b-20f10e6b9997","Type":"ContainerStarted","Data":"57950b201ca508fc0ff1d51dad68f1940a001965b8ad4cecf47be14425e6ba7f"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.102497 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adde3a48-0fe2-46e5-a82d-0ec41da371a9","Type":"ContainerDied","Data":"fc5d470725d9f6a0f985db112171be8871574d16249ca36b04a94b8e6d87d8a2"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.102540 4817 scope.go:117] "RemoveContainer" containerID="ce80e22ea9539afb7cc7dd7cd0b4aeea68fca33842a72a3c73a1262ba946fd63" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.102632 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.112212 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c768eb61-2c73-41c4-a06c-88893fdb33e9","Type":"ContainerStarted","Data":"4ed6f5dc42f27ac659fe65b98511672b27bb168a6d1239a3314baccfc460db97"} Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.112324 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.123837 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-84d84dc879-8hskz" podStartSLOduration=3.576281661 podStartE2EDuration="7.12381788s" podCreationTimestamp="2025-11-28 14:47:33 +0000 UTC" firstStartedPulling="2025-11-28 14:47:35.232534675 +0000 UTC m=+1077.820512941" lastFinishedPulling="2025-11-28 14:47:38.780070894 +0000 UTC m=+1081.368049160" observedRunningTime="2025-11-28 14:47:40.11136344 +0000 UTC m=+1082.699341706" watchObservedRunningTime="2025-11-28 14:47:40.12381788 +0000 UTC m=+1082.711796146" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.141428 4817 scope.go:117] "RemoveContainer" containerID="b93b7e5375b5813689ae7b329737e52a9e07bf7b0cc619d366018dd0a4745064" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.142261 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7dd97866bb-qxt2v" podStartSLOduration=3.779492559 podStartE2EDuration="7.142243888s" podCreationTimestamp="2025-11-28 14:47:33 +0000 UTC" firstStartedPulling="2025-11-28 14:47:35.423628442 +0000 UTC m=+1078.011606708" lastFinishedPulling="2025-11-28 14:47:38.786379761 +0000 UTC m=+1081.374358037" observedRunningTime="2025-11-28 14:47:40.140501454 +0000 UTC m=+1082.728479720" watchObservedRunningTime="2025-11-28 14:47:40.142243888 +0000 UTC m=+1082.730222154" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.168843 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" podStartSLOduration=7.168825448 podStartE2EDuration="7.168825448s" podCreationTimestamp="2025-11-28 14:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:40.164976572 +0000 UTC m=+1082.752954838" watchObservedRunningTime="2025-11-28 14:47:40.168825448 +0000 UTC m=+1082.756803714" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.187683 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.207325 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.233913 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.253679 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.299056 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: W1128 14:47:40.302039 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15698a16_9f4e_4552_b729_27969466063d.slice/crio-a428bb111e2908292054bba6443608a549361e4b64202fcb32a2a81d130c17d9 WatchSource:0}: Error finding container a428bb111e2908292054bba6443608a549361e4b64202fcb32a2a81d130c17d9: Status 404 returned error can't find the container with id a428bb111e2908292054bba6443608a549361e4b64202fcb32a2a81d130c17d9 Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.307331 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.307547 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.316859 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.317299 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.352796 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.365541 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.372347 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.164643679 podStartE2EDuration="8.372327525s" podCreationTimestamp="2025-11-28 14:47:32 +0000 UTC" firstStartedPulling="2025-11-28 14:47:35.214482086 +0000 UTC m=+1077.802460352" lastFinishedPulling="2025-11-28 14:47:36.422165932 +0000 UTC m=+1079.010144198" observedRunningTime="2025-11-28 14:47:40.265807938 +0000 UTC m=+1082.853786204" watchObservedRunningTime="2025-11-28 14:47:40.372327525 +0000 UTC m=+1082.960305791" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.373279 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.376154 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.376194 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.391547 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.402040 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fb9577b84-z8dsl"] Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.452640 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.452893 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453009 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-scripts\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453105 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ssv5\" (UniqueName: \"kubernetes.io/projected/af80cea2-4933-47de-9081-087c79157901-kube-api-access-6ssv5\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453219 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-logs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453288 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd6s2\" (UniqueName: \"kubernetes.io/projected/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-kube-api-access-qd6s2\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453405 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453496 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-config-data\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453610 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453736 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-log-httpd\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453834 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-scripts\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453900 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.453986 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-run-httpd\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.454087 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.454221 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.454294 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-config-data\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.556580 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.556655 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.556711 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-scripts\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.556755 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ssv5\" (UniqueName: \"kubernetes.io/projected/af80cea2-4933-47de-9081-087c79157901-kube-api-access-6ssv5\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.557001 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-logs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.557241 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd6s2\" (UniqueName: \"kubernetes.io/projected/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-kube-api-access-qd6s2\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.557451 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.557627 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-config-data\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.557804 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.558291 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-log-httpd\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.558450 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-scripts\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.558487 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.558690 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-run-httpd\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.559122 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.559163 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.559158 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-logs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.559415 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-log-httpd\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.559464 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-config-data\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.559916 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-run-httpd\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.560075 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.565615 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.566925 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.567763 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.568569 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.571125 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.572693 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.573003 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-scripts\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.573522 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-scripts\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.575917 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-config-data\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.579407 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-config-data\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.583654 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ssv5\" (UniqueName: \"kubernetes.io/projected/af80cea2-4933-47de-9081-087c79157901-kube-api-access-6ssv5\") pod \"ceilometer-0\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " pod="openstack/ceilometer-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.585844 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd6s2\" (UniqueName: \"kubernetes.io/projected/f8b2cd5c-abbf-4cca-98d0-e0f3db266421-kube-api-access-qd6s2\") pod \"cinder-api-0\" (UID: \"f8b2cd5c-abbf-4cca-98d0-e0f3db266421\") " pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.660794 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 14:47:40 crc kubenswrapper[4817]: I1128 14:47:40.689271 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.134269 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fb9577b84-z8dsl" event={"ID":"15698a16-9f4e-4552-b729-27969466063d","Type":"ContainerStarted","Data":"8a6a9192cfbca6714ae93ab0eb06eeca55b401aee3ecabd5cc50eaea8b6dbd4b"} Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.134607 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fb9577b84-z8dsl" event={"ID":"15698a16-9f4e-4552-b729-27969466063d","Type":"ContainerStarted","Data":"c3f2f41d28db7a7f7c1058f4360b35b2a07364af3b7af85e41f43b67361e1952"} Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.134619 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fb9577b84-z8dsl" event={"ID":"15698a16-9f4e-4552-b729-27969466063d","Type":"ContainerStarted","Data":"a428bb111e2908292054bba6443608a549361e4b64202fcb32a2a81d130c17d9"} Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.134827 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.134864 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.165741 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7fb9577b84-z8dsl" podStartSLOduration=2.165691815 podStartE2EDuration="2.165691815s" podCreationTimestamp="2025-11-28 14:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:41.153636596 +0000 UTC m=+1083.741614862" watchObservedRunningTime="2025-11-28 14:47:41.165691815 +0000 UTC m=+1083.753670081" Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.193570 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.261106 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:47:41 crc kubenswrapper[4817]: W1128 14:47:41.272986 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf80cea2_4933_47de_9081_087c79157901.slice/crio-dbc5e144ff95ebfee9f0336ddd7bf3f77aede58a5ea5fc0b268d943ddfd3eed8 WatchSource:0}: Error finding container dbc5e144ff95ebfee9f0336ddd7bf3f77aede58a5ea5fc0b268d943ddfd3eed8: Status 404 returned error can't find the container with id dbc5e144ff95ebfee9f0336ddd7bf3f77aede58a5ea5fc0b268d943ddfd3eed8 Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.747326 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49a012fe-0369-4c46-89e8-0c0062f46c33" path="/var/lib/kubelet/pods/49a012fe-0369-4c46-89e8-0c0062f46c33/volumes" Nov 28 14:47:41 crc kubenswrapper[4817]: I1128 14:47:41.749323 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adde3a48-0fe2-46e5-a82d-0ec41da371a9" path="/var/lib/kubelet/pods/adde3a48-0fe2-46e5-a82d-0ec41da371a9/volumes" Nov 28 14:47:42 crc kubenswrapper[4817]: I1128 14:47:42.156044 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f8b2cd5c-abbf-4cca-98d0-e0f3db266421","Type":"ContainerStarted","Data":"a46d31609e7d8a64361f4ce00c377a1739948721a61b68c48fca8eb2a641f05b"} Nov 28 14:47:42 crc kubenswrapper[4817]: I1128 14:47:42.156100 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f8b2cd5c-abbf-4cca-98d0-e0f3db266421","Type":"ContainerStarted","Data":"4e3fa2ce0e40e703ee082e55a8c1dd9fa05399b9ab9e14e43c8f9db6aff0521b"} Nov 28 14:47:42 crc kubenswrapper[4817]: I1128 14:47:42.158786 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerStarted","Data":"9efe7ee91cbebeea8b35d381d31947af9918e5f3f82b51e2f6d84d3117b709de"} Nov 28 14:47:42 crc kubenswrapper[4817]: I1128 14:47:42.158837 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerStarted","Data":"dbc5e144ff95ebfee9f0336ddd7bf3f77aede58a5ea5fc0b268d943ddfd3eed8"} Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.175056 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerStarted","Data":"841b4e7ef84aee9eaf39efacd49ce3af8490a33d218ffe5df30cbf35f7b19bd9"} Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.176913 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f8b2cd5c-abbf-4cca-98d0-e0f3db266421","Type":"ContainerStarted","Data":"f4d871110866e64ac6293001bbbd345b1d867280b116622c54d1c815bd0e3c23"} Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.177124 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.226144 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.226122028 podStartE2EDuration="3.226122028s" podCreationTimestamp="2025-11-28 14:47:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:43.20812362 +0000 UTC m=+1085.796101886" watchObservedRunningTime="2025-11-28 14:47:43.226122028 +0000 UTC m=+1085.814100294" Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.234164 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.444758 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.634798 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:47:43 crc kubenswrapper[4817]: I1128 14:47:43.704228 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:47:44 crc kubenswrapper[4817]: I1128 14:47:44.189488 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerStarted","Data":"e9806121c579f715bd8c685a0d3583b59f4fda6d031e834f2844c38ba5efb2d7"} Nov 28 14:47:44 crc kubenswrapper[4817]: I1128 14:47:44.194887 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:47:44 crc kubenswrapper[4817]: I1128 14:47:44.238149 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:44 crc kubenswrapper[4817]: I1128 14:47:44.300742 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-c2qft"] Nov 28 14:47:44 crc kubenswrapper[4817]: I1128 14:47:44.301015 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" containerName="dnsmasq-dns" containerID="cri-o://a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c" gracePeriod=10 Nov 28 14:47:44 crc kubenswrapper[4817]: I1128 14:47:44.887053 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gk926"] Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.017689 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.177459 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-swift-storage-0\") pod \"67758bba-e1c5-4315-8944-b1d247f2f422\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.177544 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-nb\") pod \"67758bba-e1c5-4315-8944-b1d247f2f422\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.177683 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-svc\") pod \"67758bba-e1c5-4315-8944-b1d247f2f422\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.177744 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8hf7\" (UniqueName: \"kubernetes.io/projected/67758bba-e1c5-4315-8944-b1d247f2f422-kube-api-access-x8hf7\") pod \"67758bba-e1c5-4315-8944-b1d247f2f422\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.177816 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-config\") pod \"67758bba-e1c5-4315-8944-b1d247f2f422\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.177851 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-sb\") pod \"67758bba-e1c5-4315-8944-b1d247f2f422\" (UID: \"67758bba-e1c5-4315-8944-b1d247f2f422\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.181930 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67758bba-e1c5-4315-8944-b1d247f2f422-kube-api-access-x8hf7" (OuterVolumeSpecName: "kube-api-access-x8hf7") pod "67758bba-e1c5-4315-8944-b1d247f2f422" (UID: "67758bba-e1c5-4315-8944-b1d247f2f422"). InnerVolumeSpecName "kube-api-access-x8hf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.205227 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerStarted","Data":"a819da122c263d906400cb00a7779707f3c77b4e233a15160ead325e6373d310"} Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.205604 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.208674 4817 generic.go:334] "Generic (PLEG): container finished" podID="67758bba-e1c5-4315-8944-b1d247f2f422" containerID="a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c" exitCode=0 Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.208914 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="cinder-scheduler" containerID="cri-o://bb6302042834a4d5c2a8a286e7af72073444ddb3a2ac133a8c603462430626a0" gracePeriod=30 Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.209190 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.209448 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" event={"ID":"67758bba-e1c5-4315-8944-b1d247f2f422","Type":"ContainerDied","Data":"a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c"} Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.209487 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-c2qft" event={"ID":"67758bba-e1c5-4315-8944-b1d247f2f422","Type":"ContainerDied","Data":"00ba7fbcb805553609bf843c2ee7a1af07758ab671cac5acb5ea19c5fff460db"} Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.209506 4817 scope.go:117] "RemoveContainer" containerID="a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.209710 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gk926" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" containerID="cri-o://7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e" gracePeriod=2 Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.209810 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="probe" containerID="cri-o://4ed6f5dc42f27ac659fe65b98511672b27bb168a6d1239a3314baccfc460db97" gracePeriod=30 Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.248756 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.619427739 podStartE2EDuration="5.24873669s" podCreationTimestamp="2025-11-28 14:47:40 +0000 UTC" firstStartedPulling="2025-11-28 14:47:41.275886383 +0000 UTC m=+1083.863864649" lastFinishedPulling="2025-11-28 14:47:44.905195334 +0000 UTC m=+1087.493173600" observedRunningTime="2025-11-28 14:47:45.238335801 +0000 UTC m=+1087.826314067" watchObservedRunningTime="2025-11-28 14:47:45.24873669 +0000 UTC m=+1087.836714956" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.251104 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67758bba-e1c5-4315-8944-b1d247f2f422" (UID: "67758bba-e1c5-4315-8944-b1d247f2f422"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.252136 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67758bba-e1c5-4315-8944-b1d247f2f422" (UID: "67758bba-e1c5-4315-8944-b1d247f2f422"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.253950 4817 scope.go:117] "RemoveContainer" containerID="367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.275063 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "67758bba-e1c5-4315-8944-b1d247f2f422" (UID: "67758bba-e1c5-4315-8944-b1d247f2f422"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.281393 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-config" (OuterVolumeSpecName: "config") pod "67758bba-e1c5-4315-8944-b1d247f2f422" (UID: "67758bba-e1c5-4315-8944-b1d247f2f422"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.281436 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67758bba-e1c5-4315-8944-b1d247f2f422" (UID: "67758bba-e1c5-4315-8944-b1d247f2f422"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.282072 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.282169 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.282234 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.282309 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8hf7\" (UniqueName: \"kubernetes.io/projected/67758bba-e1c5-4315-8944-b1d247f2f422-kube-api-access-x8hf7\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.290632 4817 scope.go:117] "RemoveContainer" containerID="a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c" Nov 28 14:47:45 crc kubenswrapper[4817]: E1128 14:47:45.292945 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c\": container with ID starting with a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c not found: ID does not exist" containerID="a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.293007 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c"} err="failed to get container status \"a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c\": rpc error: code = NotFound desc = could not find container \"a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c\": container with ID starting with a8f95a95cb857ee59d0c1840d6b2f7fd80a1ff134ce32808fefa6cda9ec1937c not found: ID does not exist" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.293047 4817 scope.go:117] "RemoveContainer" containerID="367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2" Nov 28 14:47:45 crc kubenswrapper[4817]: E1128 14:47:45.293511 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2\": container with ID starting with 367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2 not found: ID does not exist" containerID="367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.293539 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2"} err="failed to get container status \"367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2\": rpc error: code = NotFound desc = could not find container \"367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2\": container with ID starting with 367188649a9b892ff020330d6841e56dd2bfe3184382d3f736eb4ee20ae751c2 not found: ID does not exist" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.384009 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.384042 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67758bba-e1c5-4315-8944-b1d247f2f422-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.548392 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-c2qft"] Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.557607 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-c2qft"] Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.684402 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.747553 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" path="/var/lib/kubelet/pods/67758bba-e1c5-4315-8944-b1d247f2f422/volumes" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.796438 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt2rr\" (UniqueName: \"kubernetes.io/projected/bf21de8d-ee94-4413-b5ef-203c656d1a93-kube-api-access-kt2rr\") pod \"bf21de8d-ee94-4413-b5ef-203c656d1a93\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.796673 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-utilities\") pod \"bf21de8d-ee94-4413-b5ef-203c656d1a93\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.796759 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-catalog-content\") pod \"bf21de8d-ee94-4413-b5ef-203c656d1a93\" (UID: \"bf21de8d-ee94-4413-b5ef-203c656d1a93\") " Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.797587 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-utilities" (OuterVolumeSpecName: "utilities") pod "bf21de8d-ee94-4413-b5ef-203c656d1a93" (UID: "bf21de8d-ee94-4413-b5ef-203c656d1a93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.803754 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf21de8d-ee94-4413-b5ef-203c656d1a93-kube-api-access-kt2rr" (OuterVolumeSpecName: "kube-api-access-kt2rr") pod "bf21de8d-ee94-4413-b5ef-203c656d1a93" (UID: "bf21de8d-ee94-4413-b5ef-203c656d1a93"). InnerVolumeSpecName "kube-api-access-kt2rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.812796 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.812835 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt2rr\" (UniqueName: \"kubernetes.io/projected/bf21de8d-ee94-4413-b5ef-203c656d1a93-kube-api-access-kt2rr\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.906930 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf21de8d-ee94-4413-b5ef-203c656d1a93" (UID: "bf21de8d-ee94-4413-b5ef-203c656d1a93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:45 crc kubenswrapper[4817]: I1128 14:47:45.914741 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf21de8d-ee94-4413-b5ef-203c656d1a93-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.222627 4817 generic.go:334] "Generic (PLEG): container finished" podID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerID="4ed6f5dc42f27ac659fe65b98511672b27bb168a6d1239a3314baccfc460db97" exitCode=0 Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.222695 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c768eb61-2c73-41c4-a06c-88893fdb33e9","Type":"ContainerDied","Data":"4ed6f5dc42f27ac659fe65b98511672b27bb168a6d1239a3314baccfc460db97"} Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.225480 4817 generic.go:334] "Generic (PLEG): container finished" podID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerID="7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e" exitCode=0 Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.227399 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gk926" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.230915 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerDied","Data":"7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e"} Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.230993 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gk926" event={"ID":"bf21de8d-ee94-4413-b5ef-203c656d1a93","Type":"ContainerDied","Data":"72953d6dbe7313e15c3297aee28d2722b89bf43082793b73cbfc791738b82f9f"} Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.231020 4817 scope.go:117] "RemoveContainer" containerID="7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.268736 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.270159 4817 scope.go:117] "RemoveContainer" containerID="38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.280387 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gk926"] Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.287083 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.293806 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gk926"] Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.301544 4817 scope.go:117] "RemoveContainer" containerID="a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.346602 4817 scope.go:117] "RemoveContainer" containerID="7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e" Nov 28 14:47:46 crc kubenswrapper[4817]: E1128 14:47:46.347992 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e\": container with ID starting with 7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e not found: ID does not exist" containerID="7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.348058 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e"} err="failed to get container status \"7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e\": rpc error: code = NotFound desc = could not find container \"7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e\": container with ID starting with 7f952f037a183d69f1070668e9efbd1886e98019f8dfb2ea1cf6b5ba521b5c3e not found: ID does not exist" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.348087 4817 scope.go:117] "RemoveContainer" containerID="38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13" Nov 28 14:47:46 crc kubenswrapper[4817]: E1128 14:47:46.348696 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13\": container with ID starting with 38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13 not found: ID does not exist" containerID="38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.348741 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13"} err="failed to get container status \"38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13\": rpc error: code = NotFound desc = could not find container \"38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13\": container with ID starting with 38fea02c6a0f9703eaa2a2ae1325f6627106c563cab0ea40bf520865dae73d13 not found: ID does not exist" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.348770 4817 scope.go:117] "RemoveContainer" containerID="a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea" Nov 28 14:47:46 crc kubenswrapper[4817]: E1128 14:47:46.350601 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea\": container with ID starting with a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea not found: ID does not exist" containerID="a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea" Nov 28 14:47:46 crc kubenswrapper[4817]: I1128 14:47:46.350624 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea"} err="failed to get container status \"a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea\": rpc error: code = NotFound desc = could not find container \"a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea\": container with ID starting with a8618847fdc477b515548e010dede1f7aa042d0d27a20b490cdad156dee473ea not found: ID does not exist" Nov 28 14:47:47 crc kubenswrapper[4817]: I1128 14:47:47.751355 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" path="/var/lib/kubelet/pods/bf21de8d-ee94-4413-b5ef-203c656d1a93/volumes" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.269979 4817 generic.go:334] "Generic (PLEG): container finished" podID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerID="bb6302042834a4d5c2a8a286e7af72073444ddb3a2ac133a8c603462430626a0" exitCode=0 Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.270060 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c768eb61-2c73-41c4-a06c-88893fdb33e9","Type":"ContainerDied","Data":"bb6302042834a4d5c2a8a286e7af72073444ddb3a2ac133a8c603462430626a0"} Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.461084 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.592731 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z24k4\" (UniqueName: \"kubernetes.io/projected/c768eb61-2c73-41c4-a06c-88893fdb33e9-kube-api-access-z24k4\") pod \"c768eb61-2c73-41c4-a06c-88893fdb33e9\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.592853 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-combined-ca-bundle\") pod \"c768eb61-2c73-41c4-a06c-88893fdb33e9\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.592935 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data-custom\") pod \"c768eb61-2c73-41c4-a06c-88893fdb33e9\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.592968 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-scripts\") pod \"c768eb61-2c73-41c4-a06c-88893fdb33e9\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.593021 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c768eb61-2c73-41c4-a06c-88893fdb33e9-etc-machine-id\") pod \"c768eb61-2c73-41c4-a06c-88893fdb33e9\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.593056 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data\") pod \"c768eb61-2c73-41c4-a06c-88893fdb33e9\" (UID: \"c768eb61-2c73-41c4-a06c-88893fdb33e9\") " Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.593204 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c768eb61-2c73-41c4-a06c-88893fdb33e9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c768eb61-2c73-41c4-a06c-88893fdb33e9" (UID: "c768eb61-2c73-41c4-a06c-88893fdb33e9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.593495 4817 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c768eb61-2c73-41c4-a06c-88893fdb33e9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.599450 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-scripts" (OuterVolumeSpecName: "scripts") pod "c768eb61-2c73-41c4-a06c-88893fdb33e9" (UID: "c768eb61-2c73-41c4-a06c-88893fdb33e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.600855 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c768eb61-2c73-41c4-a06c-88893fdb33e9-kube-api-access-z24k4" (OuterVolumeSpecName: "kube-api-access-z24k4") pod "c768eb61-2c73-41c4-a06c-88893fdb33e9" (UID: "c768eb61-2c73-41c4-a06c-88893fdb33e9"). InnerVolumeSpecName "kube-api-access-z24k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.604518 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c768eb61-2c73-41c4-a06c-88893fdb33e9" (UID: "c768eb61-2c73-41c4-a06c-88893fdb33e9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.670905 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c768eb61-2c73-41c4-a06c-88893fdb33e9" (UID: "c768eb61-2c73-41c4-a06c-88893fdb33e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.697043 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.697083 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.697096 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.697110 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z24k4\" (UniqueName: \"kubernetes.io/projected/c768eb61-2c73-41c4-a06c-88893fdb33e9-kube-api-access-z24k4\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.723211 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data" (OuterVolumeSpecName: "config-data") pod "c768eb61-2c73-41c4-a06c-88893fdb33e9" (UID: "c768eb61-2c73-41c4-a06c-88893fdb33e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:49 crc kubenswrapper[4817]: I1128 14:47:49.798212 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c768eb61-2c73-41c4-a06c-88893fdb33e9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.279713 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c768eb61-2c73-41c4-a06c-88893fdb33e9","Type":"ContainerDied","Data":"2a21fd6037762090157542f448c60103103d462668ba4d10d9f735063a82a9fc"} Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.279790 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.279793 4817 scope.go:117] "RemoveContainer" containerID="4ed6f5dc42f27ac659fe65b98511672b27bb168a6d1239a3314baccfc460db97" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.305906 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.315920 4817 scope.go:117] "RemoveContainer" containerID="bb6302042834a4d5c2a8a286e7af72073444ddb3a2ac133a8c603462430626a0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.317795 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.346660 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351201 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" containerName="dnsmasq-dns" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351240 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" containerName="dnsmasq-dns" Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351288 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="extract-utilities" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351299 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="extract-utilities" Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351309 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="probe" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351319 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="probe" Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351327 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351335 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351350 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" containerName="init" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351358 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" containerName="init" Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351383 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="extract-content" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351389 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="extract-content" Nov 28 14:47:50 crc kubenswrapper[4817]: E1128 14:47:50.351401 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="cinder-scheduler" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351408 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="cinder-scheduler" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351598 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="probe" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351629 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf21de8d-ee94-4413-b5ef-203c656d1a93" containerName="registry-server" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351642 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="67758bba-e1c5-4315-8944-b1d247f2f422" containerName="dnsmasq-dns" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.351653 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" containerName="cinder-scheduler" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.352796 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.358295 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.366707 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.515978 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.516030 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-scripts\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.516053 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7570a53b-a405-45ee-ab15-ab5f575bb90b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.516267 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-config-data\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.516559 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.516635 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5nj9\" (UniqueName: \"kubernetes.io/projected/7570a53b-a405-45ee-ab15-ab5f575bb90b-kube-api-access-h5nj9\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.617937 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-scripts\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.617983 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7570a53b-a405-45ee-ab15-ab5f575bb90b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.618048 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-config-data\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.618145 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.618205 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5nj9\" (UniqueName: \"kubernetes.io/projected/7570a53b-a405-45ee-ab15-ab5f575bb90b-kube-api-access-h5nj9\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.618254 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.618650 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7570a53b-a405-45ee-ab15-ab5f575bb90b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.624631 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.626966 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-config-data\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.631169 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-scripts\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.638231 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7570a53b-a405-45ee-ab15-ab5f575bb90b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.639785 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5nj9\" (UniqueName: \"kubernetes.io/projected/7570a53b-a405-45ee-ab15-ab5f575bb90b-kube-api-access-h5nj9\") pod \"cinder-scheduler-0\" (UID: \"7570a53b-a405-45ee-ab15-ab5f575bb90b\") " pod="openstack/cinder-scheduler-0" Nov 28 14:47:50 crc kubenswrapper[4817]: I1128 14:47:50.670067 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.139440 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 14:47:51 crc kubenswrapper[4817]: W1128 14:47:51.150994 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7570a53b_a405_45ee_ab15_ab5f575bb90b.slice/crio-3499ac98233f2e7d4d214e6acc1f48a722887c3e8221838e1f3fedafcce09249 WatchSource:0}: Error finding container 3499ac98233f2e7d4d214e6acc1f48a722887c3e8221838e1f3fedafcce09249: Status 404 returned error can't find the container with id 3499ac98233f2e7d4d214e6acc1f48a722887c3e8221838e1f3fedafcce09249 Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.290963 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7570a53b-a405-45ee-ab15-ab5f575bb90b","Type":"ContainerStarted","Data":"3499ac98233f2e7d4d214e6acc1f48a722887c3e8221838e1f3fedafcce09249"} Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.327204 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.384156 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fb9577b84-z8dsl" Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.451216 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b4d5d6b8b-8n4xc"] Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.451438 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api-log" containerID="cri-o://0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c" gracePeriod=30 Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.451706 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api" containerID="cri-o://ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0" gracePeriod=30 Nov 28 14:47:51 crc kubenswrapper[4817]: I1128 14:47:51.751034 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c768eb61-2c73-41c4-a06c-88893fdb33e9" path="/var/lib/kubelet/pods/c768eb61-2c73-41c4-a06c-88893fdb33e9/volumes" Nov 28 14:47:52 crc kubenswrapper[4817]: I1128 14:47:52.306486 4817 generic.go:334] "Generic (PLEG): container finished" podID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerID="0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c" exitCode=143 Nov 28 14:47:52 crc kubenswrapper[4817]: I1128 14:47:52.306551 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" event={"ID":"137e13fa-8f5d-4076-8de7-57c8284651dd","Type":"ContainerDied","Data":"0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c"} Nov 28 14:47:52 crc kubenswrapper[4817]: I1128 14:47:52.309090 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7570a53b-a405-45ee-ab15-ab5f575bb90b","Type":"ContainerStarted","Data":"67d834b4e3b0ad6bdc361afee9641af97209b7e98f2b54a42e092b2be3ab87aa"} Nov 28 14:47:53 crc kubenswrapper[4817]: I1128 14:47:53.030285 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 28 14:47:53 crc kubenswrapper[4817]: I1128 14:47:53.316583 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7570a53b-a405-45ee-ab15-ab5f575bb90b","Type":"ContainerStarted","Data":"9ec2bf4fce480948fb4c7b12c0f9d63e14548a38e8bfc5e268fbb509c4604b2a"} Nov 28 14:47:53 crc kubenswrapper[4817]: I1128 14:47:53.345300 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.34527788 podStartE2EDuration="3.34527788s" podCreationTimestamp="2025-11-28 14:47:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:47:53.334638526 +0000 UTC m=+1095.922616792" watchObservedRunningTime="2025-11-28 14:47:53.34527788 +0000 UTC m=+1095.933256146" Nov 28 14:47:54 crc kubenswrapper[4817]: I1128 14:47:54.004438 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:54 crc kubenswrapper[4817]: I1128 14:47:54.605267 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:49952->10.217.0.163:9311: read: connection reset by peer" Nov 28 14:47:54 crc kubenswrapper[4817]: I1128 14:47:54.605277 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:49944->10.217.0.163:9311: read: connection reset by peer" Nov 28 14:47:54 crc kubenswrapper[4817]: I1128 14:47:54.827020 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c874c964f-4j6hm" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.018611 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6746568d76-zwnfd" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.084039 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.114472 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqktf\" (UniqueName: \"kubernetes.io/projected/137e13fa-8f5d-4076-8de7-57c8284651dd-kube-api-access-kqktf\") pod \"137e13fa-8f5d-4076-8de7-57c8284651dd\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.114804 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data-custom\") pod \"137e13fa-8f5d-4076-8de7-57c8284651dd\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.114976 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137e13fa-8f5d-4076-8de7-57c8284651dd-logs\") pod \"137e13fa-8f5d-4076-8de7-57c8284651dd\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.115123 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-combined-ca-bundle\") pod \"137e13fa-8f5d-4076-8de7-57c8284651dd\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.116686 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/137e13fa-8f5d-4076-8de7-57c8284651dd-logs" (OuterVolumeSpecName: "logs") pod "137e13fa-8f5d-4076-8de7-57c8284651dd" (UID: "137e13fa-8f5d-4076-8de7-57c8284651dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.126436 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "137e13fa-8f5d-4076-8de7-57c8284651dd" (UID: "137e13fa-8f5d-4076-8de7-57c8284651dd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.130022 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/137e13fa-8f5d-4076-8de7-57c8284651dd-kube-api-access-kqktf" (OuterVolumeSpecName: "kube-api-access-kqktf") pod "137e13fa-8f5d-4076-8de7-57c8284651dd" (UID: "137e13fa-8f5d-4076-8de7-57c8284651dd"). InnerVolumeSpecName "kube-api-access-kqktf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.150043 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "137e13fa-8f5d-4076-8de7-57c8284651dd" (UID: "137e13fa-8f5d-4076-8de7-57c8284651dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.216748 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data\") pod \"137e13fa-8f5d-4076-8de7-57c8284651dd\" (UID: \"137e13fa-8f5d-4076-8de7-57c8284651dd\") " Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.217248 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqktf\" (UniqueName: \"kubernetes.io/projected/137e13fa-8f5d-4076-8de7-57c8284651dd-kube-api-access-kqktf\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.217274 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.217286 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137e13fa-8f5d-4076-8de7-57c8284651dd-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.217299 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.261730 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data" (OuterVolumeSpecName: "config-data") pod "137e13fa-8f5d-4076-8de7-57c8284651dd" (UID: "137e13fa-8f5d-4076-8de7-57c8284651dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.319079 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137e13fa-8f5d-4076-8de7-57c8284651dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.337396 4817 generic.go:334] "Generic (PLEG): container finished" podID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerID="ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0" exitCode=0 Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.337447 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" event={"ID":"137e13fa-8f5d-4076-8de7-57c8284651dd","Type":"ContainerDied","Data":"ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0"} Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.337481 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" event={"ID":"137e13fa-8f5d-4076-8de7-57c8284651dd","Type":"ContainerDied","Data":"b1597ad11344b60edf23ae90f43ff30b7a52c77c7c02d2703ebb5cb5e8fecfb3"} Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.337488 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b4d5d6b8b-8n4xc" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.337505 4817 scope.go:117] "RemoveContainer" containerID="ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.376649 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b4d5d6b8b-8n4xc"] Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.379833 4817 scope.go:117] "RemoveContainer" containerID="0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.384487 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7b4d5d6b8b-8n4xc"] Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.406373 4817 scope.go:117] "RemoveContainer" containerID="ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0" Nov 28 14:47:55 crc kubenswrapper[4817]: E1128 14:47:55.406853 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0\": container with ID starting with ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0 not found: ID does not exist" containerID="ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.406890 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0"} err="failed to get container status \"ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0\": rpc error: code = NotFound desc = could not find container \"ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0\": container with ID starting with ae6eaa57a657e2bd295b1b49e0492508371268d2db6478e4bed69cbda985bad0 not found: ID does not exist" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.406917 4817 scope.go:117] "RemoveContainer" containerID="0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c" Nov 28 14:47:55 crc kubenswrapper[4817]: E1128 14:47:55.407387 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c\": container with ID starting with 0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c not found: ID does not exist" containerID="0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.407486 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c"} err="failed to get container status \"0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c\": rpc error: code = NotFound desc = could not find container \"0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c\": container with ID starting with 0eaf108f3a528b107696aa484e1785c669363692e23d66dfff9fe1e4c00c526c not found: ID does not exist" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.670950 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 14:47:55 crc kubenswrapper[4817]: I1128 14:47:55.747881 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" path="/var/lib/kubelet/pods/137e13fa-8f5d-4076-8de7-57c8284651dd/volumes" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.295600 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:56 crc kubenswrapper[4817]: E1128 14:47:56.296470 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api-log" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.296500 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api-log" Nov 28 14:47:56 crc kubenswrapper[4817]: E1128 14:47:56.296535 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.296543 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.296807 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api-log" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.296836 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="137e13fa-8f5d-4076-8de7-57c8284651dd" containerName="barbican-api" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.297602 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.299705 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-hvfzn" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.300164 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.300490 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.306989 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.335735 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.335783 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrtp4\" (UniqueName: \"kubernetes.io/projected/db5aaecd-7074-451f-bf24-a2064766b24b-kube-api-access-rrtp4\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.335948 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config-secret\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.336020 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.437210 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config-secret\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.437559 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.438233 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.439092 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.439365 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrtp4\" (UniqueName: \"kubernetes.io/projected/db5aaecd-7074-451f-bf24-a2064766b24b-kube-api-access-rrtp4\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.443259 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config-secret\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.443540 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.454550 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrtp4\" (UniqueName: \"kubernetes.io/projected/db5aaecd-7074-451f-bf24-a2064766b24b-kube-api-access-rrtp4\") pod \"openstackclient\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.615519 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.680806 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.692289 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.734688 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.736142 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.743174 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42t58\" (UniqueName: \"kubernetes.io/projected/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-kube-api-access-42t58\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.743302 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.743354 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config-secret\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.743388 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.746225 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:56 crc kubenswrapper[4817]: E1128 14:47:56.789490 4817 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 28 14:47:56 crc kubenswrapper[4817]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_db5aaecd-7074-451f-bf24-a2064766b24b_0(55939b2d007c35f7684c3bbe306956c95d8760a9c4872d7b16e4a8885d1c38c0): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"55939b2d007c35f7684c3bbe306956c95d8760a9c4872d7b16e4a8885d1c38c0" Netns:"/var/run/netns/42235663-91ba-446e-99af-56e004fe2818" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=55939b2d007c35f7684c3bbe306956c95d8760a9c4872d7b16e4a8885d1c38c0;K8S_POD_UID=db5aaecd-7074-451f-bf24-a2064766b24b" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/db5aaecd-7074-451f-bf24-a2064766b24b]: expected pod UID "db5aaecd-7074-451f-bf24-a2064766b24b" but got "dfd00f5c-51d7-4e3a-b454-bafc844d0da3" from Kube API Nov 28 14:47:56 crc kubenswrapper[4817]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 14:47:56 crc kubenswrapper[4817]: > Nov 28 14:47:56 crc kubenswrapper[4817]: E1128 14:47:56.789555 4817 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 28 14:47:56 crc kubenswrapper[4817]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_db5aaecd-7074-451f-bf24-a2064766b24b_0(55939b2d007c35f7684c3bbe306956c95d8760a9c4872d7b16e4a8885d1c38c0): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"55939b2d007c35f7684c3bbe306956c95d8760a9c4872d7b16e4a8885d1c38c0" Netns:"/var/run/netns/42235663-91ba-446e-99af-56e004fe2818" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=55939b2d007c35f7684c3bbe306956c95d8760a9c4872d7b16e4a8885d1c38c0;K8S_POD_UID=db5aaecd-7074-451f-bf24-a2064766b24b" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/db5aaecd-7074-451f-bf24-a2064766b24b]: expected pod UID "db5aaecd-7074-451f-bf24-a2064766b24b" but got "dfd00f5c-51d7-4e3a-b454-bafc844d0da3" from Kube API Nov 28 14:47:56 crc kubenswrapper[4817]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 14:47:56 crc kubenswrapper[4817]: > pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.844680 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42t58\" (UniqueName: \"kubernetes.io/projected/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-kube-api-access-42t58\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.845178 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.845346 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config-secret\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.845388 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.846368 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.854604 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.855053 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config-secret\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:56 crc kubenswrapper[4817]: I1128 14:47:56.863462 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42t58\" (UniqueName: \"kubernetes.io/projected/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-kube-api-access-42t58\") pod \"openstackclient\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " pod="openstack/openstackclient" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.067012 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.358570 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.362707 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="db5aaecd-7074-451f-bf24-a2064766b24b" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.369469 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.458246 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config-secret\") pod \"db5aaecd-7074-451f-bf24-a2064766b24b\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.458337 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-combined-ca-bundle\") pod \"db5aaecd-7074-451f-bf24-a2064766b24b\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.458416 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrtp4\" (UniqueName: \"kubernetes.io/projected/db5aaecd-7074-451f-bf24-a2064766b24b-kube-api-access-rrtp4\") pod \"db5aaecd-7074-451f-bf24-a2064766b24b\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.458564 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config\") pod \"db5aaecd-7074-451f-bf24-a2064766b24b\" (UID: \"db5aaecd-7074-451f-bf24-a2064766b24b\") " Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.459959 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "db5aaecd-7074-451f-bf24-a2064766b24b" (UID: "db5aaecd-7074-451f-bf24-a2064766b24b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.460311 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.466160 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "db5aaecd-7074-451f-bf24-a2064766b24b" (UID: "db5aaecd-7074-451f-bf24-a2064766b24b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.466290 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db5aaecd-7074-451f-bf24-a2064766b24b" (UID: "db5aaecd-7074-451f-bf24-a2064766b24b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.466688 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db5aaecd-7074-451f-bf24-a2064766b24b-kube-api-access-rrtp4" (OuterVolumeSpecName: "kube-api-access-rrtp4") pod "db5aaecd-7074-451f-bf24-a2064766b24b" (UID: "db5aaecd-7074-451f-bf24-a2064766b24b"). InnerVolumeSpecName "kube-api-access-rrtp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.512092 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 14:47:57 crc kubenswrapper[4817]: W1128 14:47:57.518535 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfd00f5c_51d7_4e3a_b454_bafc844d0da3.slice/crio-51f89fd4a2027e618ab52e15510e9bdfec7d9e115b98f86a68ef2ad745c7059c WatchSource:0}: Error finding container 51f89fd4a2027e618ab52e15510e9bdfec7d9e115b98f86a68ef2ad745c7059c: Status 404 returned error can't find the container with id 51f89fd4a2027e618ab52e15510e9bdfec7d9e115b98f86a68ef2ad745c7059c Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.562164 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.562206 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5aaecd-7074-451f-bf24-a2064766b24b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.562222 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrtp4\" (UniqueName: \"kubernetes.io/projected/db5aaecd-7074-451f-bf24-a2064766b24b-kube-api-access-rrtp4\") on node \"crc\" DevicePath \"\"" Nov 28 14:47:57 crc kubenswrapper[4817]: I1128 14:47:57.782976 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db5aaecd-7074-451f-bf24-a2064766b24b" path="/var/lib/kubelet/pods/db5aaecd-7074-451f-bf24-a2064766b24b/volumes" Nov 28 14:47:58 crc kubenswrapper[4817]: I1128 14:47:58.196559 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:47:58 crc kubenswrapper[4817]: I1128 14:47:58.374009 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dfd00f5c-51d7-4e3a-b454-bafc844d0da3","Type":"ContainerStarted","Data":"51f89fd4a2027e618ab52e15510e9bdfec7d9e115b98f86a68ef2ad745c7059c"} Nov 28 14:47:58 crc kubenswrapper[4817]: I1128 14:47:58.374127 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 14:47:58 crc kubenswrapper[4817]: I1128 14:47:58.382509 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="db5aaecd-7074-451f-bf24-a2064766b24b" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.099546 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cc96c44df-rqchh" Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.153528 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c989bbc58-5p729"] Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.155582 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c989bbc58-5p729" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-httpd" containerID="cri-o://502f3a32af05550c174a7e7b618b7b84a04508d0b5bc48c2e2d5d965bbe03808" gracePeriod=30 Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.155885 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c989bbc58-5p729" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-api" containerID="cri-o://daee428c9011ec401bff6caa7e5087bc7c3a041e8f72ad71a11316503a44d3fe" gracePeriod=30 Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.393613 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerID="502f3a32af05550c174a7e7b618b7b84a04508d0b5bc48c2e2d5d965bbe03808" exitCode=0 Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.393782 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c989bbc58-5p729" event={"ID":"eb5a94ef-ba23-48a4-99e6-23f6764f8606","Type":"ContainerDied","Data":"502f3a32af05550c174a7e7b618b7b84a04508d0b5bc48c2e2d5d965bbe03808"} Nov 28 14:48:00 crc kubenswrapper[4817]: I1128 14:48:00.909823 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.578448 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.578859 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-central-agent" containerID="cri-o://9efe7ee91cbebeea8b35d381d31947af9918e5f3f82b51e2f6d84d3117b709de" gracePeriod=30 Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.579685 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-notification-agent" containerID="cri-o://841b4e7ef84aee9eaf39efacd49ce3af8490a33d218ffe5df30cbf35f7b19bd9" gracePeriod=30 Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.579760 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="sg-core" containerID="cri-o://e9806121c579f715bd8c685a0d3583b59f4fda6d031e834f2844c38ba5efb2d7" gracePeriod=30 Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.579878 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="proxy-httpd" containerID="cri-o://a819da122c263d906400cb00a7779707f3c77b4e233a15160ead325e6373d310" gracePeriod=30 Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.586399 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.925684 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-85df6b4fc9-7c5nt"] Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.927600 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.931043 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.933368 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.933512 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.948230 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85df6b4fc9-7c5nt"] Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949523 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-internal-tls-certs\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949587 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72c26e96-3946-44a8-bb20-7786cba96b07-log-httpd\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949616 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg24k\" (UniqueName: \"kubernetes.io/projected/72c26e96-3946-44a8-bb20-7786cba96b07-kube-api-access-fg24k\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949640 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72c26e96-3946-44a8-bb20-7786cba96b07-run-httpd\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949665 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72c26e96-3946-44a8-bb20-7786cba96b07-etc-swift\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949740 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-config-data\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949792 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-public-tls-certs\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:01 crc kubenswrapper[4817]: I1128 14:48:01.949840 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-combined-ca-bundle\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052128 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-config-data\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052205 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-public-tls-certs\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052260 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-combined-ca-bundle\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052342 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-internal-tls-certs\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052381 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72c26e96-3946-44a8-bb20-7786cba96b07-log-httpd\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052413 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg24k\" (UniqueName: \"kubernetes.io/projected/72c26e96-3946-44a8-bb20-7786cba96b07-kube-api-access-fg24k\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052440 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72c26e96-3946-44a8-bb20-7786cba96b07-run-httpd\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052463 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72c26e96-3946-44a8-bb20-7786cba96b07-etc-swift\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052921 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72c26e96-3946-44a8-bb20-7786cba96b07-log-httpd\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.052961 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72c26e96-3946-44a8-bb20-7786cba96b07-run-httpd\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.059562 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-combined-ca-bundle\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.059565 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-config-data\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.060109 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-public-tls-certs\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.069431 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72c26e96-3946-44a8-bb20-7786cba96b07-etc-swift\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.070184 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg24k\" (UniqueName: \"kubernetes.io/projected/72c26e96-3946-44a8-bb20-7786cba96b07-kube-api-access-fg24k\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.072362 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c26e96-3946-44a8-bb20-7786cba96b07-internal-tls-certs\") pod \"swift-proxy-85df6b4fc9-7c5nt\" (UID: \"72c26e96-3946-44a8-bb20-7786cba96b07\") " pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.266530 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.430081 4817 generic.go:334] "Generic (PLEG): container finished" podID="af80cea2-4933-47de-9081-087c79157901" containerID="a819da122c263d906400cb00a7779707f3c77b4e233a15160ead325e6373d310" exitCode=0 Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.430115 4817 generic.go:334] "Generic (PLEG): container finished" podID="af80cea2-4933-47de-9081-087c79157901" containerID="e9806121c579f715bd8c685a0d3583b59f4fda6d031e834f2844c38ba5efb2d7" exitCode=2 Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.430122 4817 generic.go:334] "Generic (PLEG): container finished" podID="af80cea2-4933-47de-9081-087c79157901" containerID="9efe7ee91cbebeea8b35d381d31947af9918e5f3f82b51e2f6d84d3117b709de" exitCode=0 Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.430120 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerDied","Data":"a819da122c263d906400cb00a7779707f3c77b4e233a15160ead325e6373d310"} Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.430167 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerDied","Data":"e9806121c579f715bd8c685a0d3583b59f4fda6d031e834f2844c38ba5efb2d7"} Nov 28 14:48:02 crc kubenswrapper[4817]: I1128 14:48:02.430177 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerDied","Data":"9efe7ee91cbebeea8b35d381d31947af9918e5f3f82b51e2f6d84d3117b709de"} Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.086820 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-vlrzf"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.088280 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.104623 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1324b91-71af-44bd-9658-7727c10193c6-operator-scripts\") pod \"nova-api-db-create-vlrzf\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.104686 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvtbs\" (UniqueName: \"kubernetes.io/projected/d1324b91-71af-44bd-9658-7727c10193c6-kube-api-access-lvtbs\") pod \"nova-api-db-create-vlrzf\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.104790 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vlrzf"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.185866 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-n67gd"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.187033 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.200804 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n67gd"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.205778 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1324b91-71af-44bd-9658-7727c10193c6-operator-scripts\") pod \"nova-api-db-create-vlrzf\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.205819 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knvbx\" (UniqueName: \"kubernetes.io/projected/a3f2262a-8a40-47de-9e14-d29ead210c3e-kube-api-access-knvbx\") pod \"nova-cell0-db-create-n67gd\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.205853 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvtbs\" (UniqueName: \"kubernetes.io/projected/d1324b91-71af-44bd-9658-7727c10193c6-kube-api-access-lvtbs\") pod \"nova-api-db-create-vlrzf\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.205933 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f2262a-8a40-47de-9e14-d29ead210c3e-operator-scripts\") pod \"nova-cell0-db-create-n67gd\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.207019 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1324b91-71af-44bd-9658-7727c10193c6-operator-scripts\") pod \"nova-api-db-create-vlrzf\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.232846 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvtbs\" (UniqueName: \"kubernetes.io/projected/d1324b91-71af-44bd-9658-7727c10193c6-kube-api-access-lvtbs\") pod \"nova-api-db-create-vlrzf\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.307914 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f2262a-8a40-47de-9e14-d29ead210c3e-operator-scripts\") pod \"nova-cell0-db-create-n67gd\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.308114 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knvbx\" (UniqueName: \"kubernetes.io/projected/a3f2262a-8a40-47de-9e14-d29ead210c3e-kube-api-access-knvbx\") pod \"nova-cell0-db-create-n67gd\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.308706 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f2262a-8a40-47de-9e14-d29ead210c3e-operator-scripts\") pod \"nova-cell0-db-create-n67gd\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.311426 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-063a-account-create-update-95fpn"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.326063 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-063a-account-create-update-95fpn"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.326496 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.328840 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.330744 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knvbx\" (UniqueName: \"kubernetes.io/projected/a3f2262a-8a40-47de-9e14-d29ead210c3e-kube-api-access-knvbx\") pod \"nova-cell0-db-create-n67gd\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.401126 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-62dsg"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.402581 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.410066 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fe692be-892a-46b0-b168-fa66ffce6762-operator-scripts\") pod \"nova-api-063a-account-create-update-95fpn\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.410135 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9t9d\" (UniqueName: \"kubernetes.io/projected/9fe692be-892a-46b0-b168-fa66ffce6762-kube-api-access-f9t9d\") pod \"nova-api-063a-account-create-update-95fpn\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.413933 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-62dsg"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.462734 4817 generic.go:334] "Generic (PLEG): container finished" podID="af80cea2-4933-47de-9081-087c79157901" containerID="841b4e7ef84aee9eaf39efacd49ce3af8490a33d218ffe5df30cbf35f7b19bd9" exitCode=0 Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.462745 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerDied","Data":"841b4e7ef84aee9eaf39efacd49ce3af8490a33d218ffe5df30cbf35f7b19bd9"} Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.468100 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerID="daee428c9011ec401bff6caa7e5087bc7c3a041e8f72ad71a11316503a44d3fe" exitCode=0 Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.468137 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c989bbc58-5p729" event={"ID":"eb5a94ef-ba23-48a4-99e6-23f6764f8606","Type":"ContainerDied","Data":"daee428c9011ec401bff6caa7e5087bc7c3a041e8f72ad71a11316503a44d3fe"} Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.479665 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.496929 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-16d1-account-create-update-krscn"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.498033 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.501677 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.508080 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.510927 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnhfb\" (UniqueName: \"kubernetes.io/projected/5f4757cf-18ca-406d-92ae-d975f16cb209-kube-api-access-bnhfb\") pod \"nova-cell0-16d1-account-create-update-krscn\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.510988 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fe692be-892a-46b0-b168-fa66ffce6762-operator-scripts\") pod \"nova-api-063a-account-create-update-95fpn\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.511024 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9t9d\" (UniqueName: \"kubernetes.io/projected/9fe692be-892a-46b0-b168-fa66ffce6762-kube-api-access-f9t9d\") pod \"nova-api-063a-account-create-update-95fpn\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.511088 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74102797-fa18-4bb1-a981-c5c359833031-operator-scripts\") pod \"nova-cell1-db-create-62dsg\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.511113 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f4757cf-18ca-406d-92ae-d975f16cb209-operator-scripts\") pod \"nova-cell0-16d1-account-create-update-krscn\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.511130 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjrpf\" (UniqueName: \"kubernetes.io/projected/74102797-fa18-4bb1-a981-c5c359833031-kube-api-access-cjrpf\") pod \"nova-cell1-db-create-62dsg\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.511925 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fe692be-892a-46b0-b168-fa66ffce6762-operator-scripts\") pod \"nova-api-063a-account-create-update-95fpn\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.517502 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-16d1-account-create-update-krscn"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.534873 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9t9d\" (UniqueName: \"kubernetes.io/projected/9fe692be-892a-46b0-b168-fa66ffce6762-kube-api-access-f9t9d\") pod \"nova-api-063a-account-create-update-95fpn\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.612305 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74102797-fa18-4bb1-a981-c5c359833031-operator-scripts\") pod \"nova-cell1-db-create-62dsg\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.612579 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f4757cf-18ca-406d-92ae-d975f16cb209-operator-scripts\") pod \"nova-cell0-16d1-account-create-update-krscn\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.612597 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjrpf\" (UniqueName: \"kubernetes.io/projected/74102797-fa18-4bb1-a981-c5c359833031-kube-api-access-cjrpf\") pod \"nova-cell1-db-create-62dsg\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.612644 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnhfb\" (UniqueName: \"kubernetes.io/projected/5f4757cf-18ca-406d-92ae-d975f16cb209-kube-api-access-bnhfb\") pod \"nova-cell0-16d1-account-create-update-krscn\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.613468 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74102797-fa18-4bb1-a981-c5c359833031-operator-scripts\") pod \"nova-cell1-db-create-62dsg\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.614029 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f4757cf-18ca-406d-92ae-d975f16cb209-operator-scripts\") pod \"nova-cell0-16d1-account-create-update-krscn\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.629366 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnhfb\" (UniqueName: \"kubernetes.io/projected/5f4757cf-18ca-406d-92ae-d975f16cb209-kube-api-access-bnhfb\") pod \"nova-cell0-16d1-account-create-update-krscn\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.632065 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjrpf\" (UniqueName: \"kubernetes.io/projected/74102797-fa18-4bb1-a981-c5c359833031-kube-api-access-cjrpf\") pod \"nova-cell1-db-create-62dsg\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.701786 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d5f5-account-create-update-2hjmb"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.703181 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.705632 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.713457 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d5f5-account-create-update-2hjmb"] Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.715802 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn42q\" (UniqueName: \"kubernetes.io/projected/6afd8c02-96a9-44c5-a773-a137ab3df6e9-kube-api-access-xn42q\") pod \"nova-cell1-d5f5-account-create-update-2hjmb\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.715870 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6afd8c02-96a9-44c5-a773-a137ab3df6e9-operator-scripts\") pod \"nova-cell1-d5f5-account-create-update-2hjmb\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.727067 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.733700 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.817713 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn42q\" (UniqueName: \"kubernetes.io/projected/6afd8c02-96a9-44c5-a773-a137ab3df6e9-kube-api-access-xn42q\") pod \"nova-cell1-d5f5-account-create-update-2hjmb\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.817907 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6afd8c02-96a9-44c5-a773-a137ab3df6e9-operator-scripts\") pod \"nova-cell1-d5f5-account-create-update-2hjmb\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.818882 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6afd8c02-96a9-44c5-a773-a137ab3df6e9-operator-scripts\") pod \"nova-cell1-d5f5-account-create-update-2hjmb\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.839021 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn42q\" (UniqueName: \"kubernetes.io/projected/6afd8c02-96a9-44c5-a773-a137ab3df6e9-kube-api-access-xn42q\") pod \"nova-cell1-d5f5-account-create-update-2hjmb\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:05 crc kubenswrapper[4817]: I1128 14:48:05.919139 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:06 crc kubenswrapper[4817]: I1128 14:48:06.029066 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.646745 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.647452 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-log" containerID="cri-o://6fa052b7b40112b5990ebbff5d9a4b45c427e0cab7a5e5ecd75d6779e0bb83da" gracePeriod=30 Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.648019 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-httpd" containerID="cri-o://f4f22ca1b0a23418a0255001362927ff0e4b8f602cb4785a1133d170f894b648" gracePeriod=30 Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.814164 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.984669 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ssv5\" (UniqueName: \"kubernetes.io/projected/af80cea2-4933-47de-9081-087c79157901-kube-api-access-6ssv5\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.984785 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-combined-ca-bundle\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.984909 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-run-httpd\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.984931 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-scripts\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.984985 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-sg-core-conf-yaml\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.984999 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-config-data\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.985051 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-log-httpd\") pod \"af80cea2-4933-47de-9081-087c79157901\" (UID: \"af80cea2-4933-47de-9081-087c79157901\") " Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.986578 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.986639 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:07 crc kubenswrapper[4817]: I1128 14:48:07.993189 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af80cea2-4933-47de-9081-087c79157901-kube-api-access-6ssv5" (OuterVolumeSpecName: "kube-api-access-6ssv5") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "kube-api-access-6ssv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:07.999391 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-scripts" (OuterVolumeSpecName: "scripts") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.030049 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.087838 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.087880 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ssv5\" (UniqueName: \"kubernetes.io/projected/af80cea2-4933-47de-9081-087c79157901-kube-api-access-6ssv5\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.087892 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af80cea2-4933-47de-9081-087c79157901-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.087900 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.098942 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.155124 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.191303 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-ovndb-tls-certs\") pod \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.191415 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-httpd-config\") pod \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.191485 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-combined-ca-bundle\") pod \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.191516 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9hqj\" (UniqueName: \"kubernetes.io/projected/eb5a94ef-ba23-48a4-99e6-23f6764f8606-kube-api-access-f9hqj\") pod \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.191993 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-config\") pod \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\" (UID: \"eb5a94ef-ba23-48a4-99e6-23f6764f8606\") " Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.192470 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.192486 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.194070 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb5a94ef-ba23-48a4-99e6-23f6764f8606-kube-api-access-f9hqj" (OuterVolumeSpecName: "kube-api-access-f9hqj") pod "eb5a94ef-ba23-48a4-99e6-23f6764f8606" (UID: "eb5a94ef-ba23-48a4-99e6-23f6764f8606"). InnerVolumeSpecName "kube-api-access-f9hqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.197363 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "eb5a94ef-ba23-48a4-99e6-23f6764f8606" (UID: "eb5a94ef-ba23-48a4-99e6-23f6764f8606"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.249938 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-config-data" (OuterVolumeSpecName: "config-data") pod "af80cea2-4933-47de-9081-087c79157901" (UID: "af80cea2-4933-47de-9081-087c79157901"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.259345 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb5a94ef-ba23-48a4-99e6-23f6764f8606" (UID: "eb5a94ef-ba23-48a4-99e6-23f6764f8606"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.285743 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "eb5a94ef-ba23-48a4-99e6-23f6764f8606" (UID: "eb5a94ef-ba23-48a4-99e6-23f6764f8606"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.288181 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-config" (OuterVolumeSpecName: "config") pod "eb5a94ef-ba23-48a4-99e6-23f6764f8606" (UID: "eb5a94ef-ba23-48a4-99e6-23f6764f8606"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.296418 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af80cea2-4933-47de-9081-087c79157901-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.296480 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.296498 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9hqj\" (UniqueName: \"kubernetes.io/projected/eb5a94ef-ba23-48a4-99e6-23f6764f8606-kube-api-access-f9hqj\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.296511 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.296522 4817 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.296532 4817 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eb5a94ef-ba23-48a4-99e6-23f6764f8606-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.441983 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-62dsg"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.485370 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n67gd"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.516131 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85df6b4fc9-7c5nt"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.527753 4817 generic.go:334] "Generic (PLEG): container finished" podID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerID="6fa052b7b40112b5990ebbff5d9a4b45c427e0cab7a5e5ecd75d6779e0bb83da" exitCode=143 Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.527827 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9f53839-25c4-4e8f-b9c7-1895ddff10d6","Type":"ContainerDied","Data":"6fa052b7b40112b5990ebbff5d9a4b45c427e0cab7a5e5ecd75d6779e0bb83da"} Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.532390 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n67gd" event={"ID":"a3f2262a-8a40-47de-9e14-d29ead210c3e","Type":"ContainerStarted","Data":"9511b9f8560cd698602c6682b1df9ac92dcb592ac1715f4c7a4548f78df1c1c9"} Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.537222 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c989bbc58-5p729" event={"ID":"eb5a94ef-ba23-48a4-99e6-23f6764f8606","Type":"ContainerDied","Data":"cd0eea617e090630850a3e62e6a41d2477bc998c1709306bc10579df914cf233"} Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.537284 4817 scope.go:117] "RemoveContainer" containerID="502f3a32af05550c174a7e7b618b7b84a04508d0b5bc48c2e2d5d965bbe03808" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.537441 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c989bbc58-5p729" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.540971 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dfd00f5c-51d7-4e3a-b454-bafc844d0da3","Type":"ContainerStarted","Data":"99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa"} Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.546749 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af80cea2-4933-47de-9081-087c79157901","Type":"ContainerDied","Data":"dbc5e144ff95ebfee9f0336ddd7bf3f77aede58a5ea5fc0b268d943ddfd3eed8"} Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.546844 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.560942 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.618106086 podStartE2EDuration="12.560928087s" podCreationTimestamp="2025-11-28 14:47:56 +0000 UTC" firstStartedPulling="2025-11-28 14:47:57.521661024 +0000 UTC m=+1100.109639290" lastFinishedPulling="2025-11-28 14:48:07.464483035 +0000 UTC m=+1110.052461291" observedRunningTime="2025-11-28 14:48:08.554168149 +0000 UTC m=+1111.142146415" watchObservedRunningTime="2025-11-28 14:48:08.560928087 +0000 UTC m=+1111.148906353" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.562294 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-62dsg" event={"ID":"74102797-fa18-4bb1-a981-c5c359833031","Type":"ContainerStarted","Data":"7d74feff7b9b2c3bdc2dc65b1c9f287557bcca04f5abf6ee60041bad74949c01"} Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.625606 4817 scope.go:117] "RemoveContainer" containerID="daee428c9011ec401bff6caa7e5087bc7c3a041e8f72ad71a11316503a44d3fe" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.634865 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-063a-account-create-update-95fpn"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.649366 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d5f5-account-create-update-2hjmb"] Nov 28 14:48:08 crc kubenswrapper[4817]: W1128 14:48:08.650832 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fe692be_892a_46b0_b168_fa66ffce6762.slice/crio-c5f7dc477d522788edfbc80719cd75b2967b3358284e67de36c8dbcddf4bfb53 WatchSource:0}: Error finding container c5f7dc477d522788edfbc80719cd75b2967b3358284e67de36c8dbcddf4bfb53: Status 404 returned error can't find the container with id c5f7dc477d522788edfbc80719cd75b2967b3358284e67de36c8dbcddf4bfb53 Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.662643 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-16d1-account-create-update-krscn"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.681624 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c989bbc58-5p729"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.693877 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vlrzf"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.699157 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c989bbc58-5p729"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.712625 4817 scope.go:117] "RemoveContainer" containerID="a819da122c263d906400cb00a7779707f3c77b4e233a15160ead325e6373d310" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.717506 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.726736 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.743860 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:08 crc kubenswrapper[4817]: E1128 14:48:08.744233 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-central-agent" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744248 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-central-agent" Nov 28 14:48:08 crc kubenswrapper[4817]: E1128 14:48:08.744263 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-notification-agent" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744269 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-notification-agent" Nov 28 14:48:08 crc kubenswrapper[4817]: E1128 14:48:08.744286 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-api" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744292 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-api" Nov 28 14:48:08 crc kubenswrapper[4817]: E1128 14:48:08.744305 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="proxy-httpd" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744310 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="proxy-httpd" Nov 28 14:48:08 crc kubenswrapper[4817]: E1128 14:48:08.744324 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-httpd" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744330 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-httpd" Nov 28 14:48:08 crc kubenswrapper[4817]: E1128 14:48:08.744344 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="sg-core" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744349 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="sg-core" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744569 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="sg-core" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744584 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="proxy-httpd" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744598 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-notification-agent" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744609 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="af80cea2-4933-47de-9081-087c79157901" containerName="ceilometer-central-agent" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744618 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-httpd" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.744628 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" containerName="neutron-api" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.746274 4817 scope.go:117] "RemoveContainer" containerID="e9806121c579f715bd8c685a0d3583b59f4fda6d031e834f2844c38ba5efb2d7" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.747301 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.750238 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.750489 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.759182 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.798823 4817 scope.go:117] "RemoveContainer" containerID="841b4e7ef84aee9eaf39efacd49ce3af8490a33d218ffe5df30cbf35f7b19bd9" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.858528 4817 scope.go:117] "RemoveContainer" containerID="9efe7ee91cbebeea8b35d381d31947af9918e5f3f82b51e2f6d84d3117b709de" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921002 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921105 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-log-httpd\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921133 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-config-data\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921171 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-scripts\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921221 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921253 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2gh8\" (UniqueName: \"kubernetes.io/projected/eb11a253-9431-42b9-8d11-ad971f486253-kube-api-access-b2gh8\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:08 crc kubenswrapper[4817]: I1128 14:48:08.921335 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-run-httpd\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.024898 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-run-httpd\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.024980 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.025045 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-log-httpd\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.025066 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-config-data\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.025096 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-scripts\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.025131 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.025157 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2gh8\" (UniqueName: \"kubernetes.io/projected/eb11a253-9431-42b9-8d11-ad971f486253-kube-api-access-b2gh8\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.025870 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-log-httpd\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.026194 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-run-httpd\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.031210 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.034366 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-config-data\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.034383 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-scripts\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.037972 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.051527 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2gh8\" (UniqueName: \"kubernetes.io/projected/eb11a253-9431-42b9-8d11-ad971f486253-kube-api-access-b2gh8\") pod \"ceilometer-0\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.148915 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.588375 4817 generic.go:334] "Generic (PLEG): container finished" podID="74102797-fa18-4bb1-a981-c5c359833031" containerID="9ebe3eab7e0802348099cb78a107daa6c1e074795a2cce2c2c9c80c8ed0bfc85" exitCode=0 Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.588516 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-62dsg" event={"ID":"74102797-fa18-4bb1-a981-c5c359833031","Type":"ContainerDied","Data":"9ebe3eab7e0802348099cb78a107daa6c1e074795a2cce2c2c9c80c8ed0bfc85"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.594651 4817 generic.go:334] "Generic (PLEG): container finished" podID="a3f2262a-8a40-47de-9e14-d29ead210c3e" containerID="dba2fd87d6cef90f93fb264da5f5255b013e0b1135f974b54190c0527dafe6aa" exitCode=0 Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.594859 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n67gd" event={"ID":"a3f2262a-8a40-47de-9e14-d29ead210c3e","Type":"ContainerDied","Data":"dba2fd87d6cef90f93fb264da5f5255b013e0b1135f974b54190c0527dafe6aa"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.618879 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" event={"ID":"72c26e96-3946-44a8-bb20-7786cba96b07","Type":"ContainerStarted","Data":"e0a1fdc25bd58471988c9eec04ba493edd72bf29b860b2189185b77071bbd174"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.618919 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" event={"ID":"72c26e96-3946-44a8-bb20-7786cba96b07","Type":"ContainerStarted","Data":"daf2cbff15fe53e21f2bd0003bcf8af07d3010aeb0d8268b5d19a333906874e6"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.618929 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" event={"ID":"72c26e96-3946-44a8-bb20-7786cba96b07","Type":"ContainerStarted","Data":"9d25848b676f104fbb38e2847f12c85834bfe2b54a5e413670413708f5d6c79c"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.619429 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.625503 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.637940 4817 generic.go:334] "Generic (PLEG): container finished" podID="d1324b91-71af-44bd-9658-7727c10193c6" containerID="e484f11b74131240b7c6f0967c8b9d2cad42bbbe3950594e6af19b79cb323028" exitCode=0 Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.638067 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vlrzf" event={"ID":"d1324b91-71af-44bd-9658-7727c10193c6","Type":"ContainerDied","Data":"e484f11b74131240b7c6f0967c8b9d2cad42bbbe3950594e6af19b79cb323028"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.638105 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vlrzf" event={"ID":"d1324b91-71af-44bd-9658-7727c10193c6","Type":"ContainerStarted","Data":"945f3ac0f567a0db646f762c9b69485a02b31553b637377ffb61a85090c1801d"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.647951 4817 generic.go:334] "Generic (PLEG): container finished" podID="9fe692be-892a-46b0-b168-fa66ffce6762" containerID="b1a5a935c61b98455ac4ca92ef26ac7fc0a899197d1b85ed577e0df4713a3cde" exitCode=0 Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.648114 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-063a-account-create-update-95fpn" event={"ID":"9fe692be-892a-46b0-b168-fa66ffce6762","Type":"ContainerDied","Data":"b1a5a935c61b98455ac4ca92ef26ac7fc0a899197d1b85ed577e0df4713a3cde"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.648156 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-063a-account-create-update-95fpn" event={"ID":"9fe692be-892a-46b0-b168-fa66ffce6762","Type":"ContainerStarted","Data":"c5f7dc477d522788edfbc80719cd75b2967b3358284e67de36c8dbcddf4bfb53"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.655482 4817 generic.go:334] "Generic (PLEG): container finished" podID="5f4757cf-18ca-406d-92ae-d975f16cb209" containerID="c212fe10c7fd47f48512470d10d45586c44e5d0bb639d9d009663532bb88c14a" exitCode=0 Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.655571 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-16d1-account-create-update-krscn" event={"ID":"5f4757cf-18ca-406d-92ae-d975f16cb209","Type":"ContainerDied","Data":"c212fe10c7fd47f48512470d10d45586c44e5d0bb639d9d009663532bb88c14a"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.655603 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-16d1-account-create-update-krscn" event={"ID":"5f4757cf-18ca-406d-92ae-d975f16cb209","Type":"ContainerStarted","Data":"8ec768370a3a4e51083cfff7a57dcea48fba4eb0a12c5320ce02bc4bfab4bd72"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.683879 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.684858 4817 generic.go:334] "Generic (PLEG): container finished" podID="6afd8c02-96a9-44c5-a773-a137ab3df6e9" containerID="06850586edffc8451a9e554b885a2782a2b8c9d6af6cc5b5ec8d9a4dfea1c71e" exitCode=0 Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.685463 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" event={"ID":"6afd8c02-96a9-44c5-a773-a137ab3df6e9","Type":"ContainerDied","Data":"06850586edffc8451a9e554b885a2782a2b8c9d6af6cc5b5ec8d9a4dfea1c71e"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.685510 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" event={"ID":"6afd8c02-96a9-44c5-a773-a137ab3df6e9","Type":"ContainerStarted","Data":"84dbf4efaf3022e2d61704adfe7f47a267fd224bf2786da9dfaaa0ad980e7182"} Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.714139 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" podStartSLOduration=8.714112017 podStartE2EDuration="8.714112017s" podCreationTimestamp="2025-11-28 14:48:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:09.667495369 +0000 UTC m=+1112.255473635" watchObservedRunningTime="2025-11-28 14:48:09.714112017 +0000 UTC m=+1112.302090293" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.750051 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af80cea2-4933-47de-9081-087c79157901" path="/var/lib/kubelet/pods/af80cea2-4933-47de-9081-087c79157901/volumes" Nov 28 14:48:09 crc kubenswrapper[4817]: I1128 14:48:09.751590 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb5a94ef-ba23-48a4-99e6-23f6764f8606" path="/var/lib/kubelet/pods/eb5a94ef-ba23-48a4-99e6-23f6764f8606/volumes" Nov 28 14:48:10 crc kubenswrapper[4817]: I1128 14:48:10.699186 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerStarted","Data":"de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948"} Nov 28 14:48:10 crc kubenswrapper[4817]: I1128 14:48:10.699795 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerStarted","Data":"f7664cce90990e38d44a9be2ebf22249ef872088adac1d55790bfd15ddfb3439"} Nov 28 14:48:10 crc kubenswrapper[4817]: I1128 14:48:10.764556 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:48:10 crc kubenswrapper[4817]: I1128 14:48:10.764878 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-log" containerID="cri-o://57478e58c4952c9a190ce4ef2fc26d7345be350a1e244deee775c1d100af9f6a" gracePeriod=30 Nov 28 14:48:10 crc kubenswrapper[4817]: I1128 14:48:10.764976 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-httpd" containerID="cri-o://7d6ec80e36c3086c48828d66ba9659a186575524e70c5be9355705832797e3d6" gracePeriod=30 Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.152276 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.283529 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.311579 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6afd8c02-96a9-44c5-a773-a137ab3df6e9-operator-scripts\") pod \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.311624 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn42q\" (UniqueName: \"kubernetes.io/projected/6afd8c02-96a9-44c5-a773-a137ab3df6e9-kube-api-access-xn42q\") pod \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\" (UID: \"6afd8c02-96a9-44c5-a773-a137ab3df6e9\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.312330 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6afd8c02-96a9-44c5-a773-a137ab3df6e9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6afd8c02-96a9-44c5-a773-a137ab3df6e9" (UID: "6afd8c02-96a9-44c5-a773-a137ab3df6e9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.324923 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6afd8c02-96a9-44c5-a773-a137ab3df6e9-kube-api-access-xn42q" (OuterVolumeSpecName: "kube-api-access-xn42q") pod "6afd8c02-96a9-44c5-a773-a137ab3df6e9" (UID: "6afd8c02-96a9-44c5-a773-a137ab3df6e9"). InnerVolumeSpecName "kube-api-access-xn42q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.414200 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6afd8c02-96a9-44c5-a773-a137ab3df6e9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.414230 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn42q\" (UniqueName: \"kubernetes.io/projected/6afd8c02-96a9-44c5-a773-a137ab3df6e9-kube-api-access-xn42q\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.491058 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.502522 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.511099 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.513222 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.524245 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617104 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9t9d\" (UniqueName: \"kubernetes.io/projected/9fe692be-892a-46b0-b168-fa66ffce6762-kube-api-access-f9t9d\") pod \"9fe692be-892a-46b0-b168-fa66ffce6762\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617480 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvtbs\" (UniqueName: \"kubernetes.io/projected/d1324b91-71af-44bd-9658-7727c10193c6-kube-api-access-lvtbs\") pod \"d1324b91-71af-44bd-9658-7727c10193c6\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617569 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjrpf\" (UniqueName: \"kubernetes.io/projected/74102797-fa18-4bb1-a981-c5c359833031-kube-api-access-cjrpf\") pod \"74102797-fa18-4bb1-a981-c5c359833031\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617595 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74102797-fa18-4bb1-a981-c5c359833031-operator-scripts\") pod \"74102797-fa18-4bb1-a981-c5c359833031\" (UID: \"74102797-fa18-4bb1-a981-c5c359833031\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617671 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1324b91-71af-44bd-9658-7727c10193c6-operator-scripts\") pod \"d1324b91-71af-44bd-9658-7727c10193c6\" (UID: \"d1324b91-71af-44bd-9658-7727c10193c6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617735 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f4757cf-18ca-406d-92ae-d975f16cb209-operator-scripts\") pod \"5f4757cf-18ca-406d-92ae-d975f16cb209\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617756 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnhfb\" (UniqueName: \"kubernetes.io/projected/5f4757cf-18ca-406d-92ae-d975f16cb209-kube-api-access-bnhfb\") pod \"5f4757cf-18ca-406d-92ae-d975f16cb209\" (UID: \"5f4757cf-18ca-406d-92ae-d975f16cb209\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.617774 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fe692be-892a-46b0-b168-fa66ffce6762-operator-scripts\") pod \"9fe692be-892a-46b0-b168-fa66ffce6762\" (UID: \"9fe692be-892a-46b0-b168-fa66ffce6762\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.618421 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74102797-fa18-4bb1-a981-c5c359833031-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "74102797-fa18-4bb1-a981-c5c359833031" (UID: "74102797-fa18-4bb1-a981-c5c359833031"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.618551 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fe692be-892a-46b0-b168-fa66ffce6762-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9fe692be-892a-46b0-b168-fa66ffce6762" (UID: "9fe692be-892a-46b0-b168-fa66ffce6762"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.618914 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1324b91-71af-44bd-9658-7727c10193c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1324b91-71af-44bd-9658-7727c10193c6" (UID: "d1324b91-71af-44bd-9658-7727c10193c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.619282 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f4757cf-18ca-406d-92ae-d975f16cb209-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f4757cf-18ca-406d-92ae-d975f16cb209" (UID: "5f4757cf-18ca-406d-92ae-d975f16cb209"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.622099 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74102797-fa18-4bb1-a981-c5c359833031-kube-api-access-cjrpf" (OuterVolumeSpecName: "kube-api-access-cjrpf") pod "74102797-fa18-4bb1-a981-c5c359833031" (UID: "74102797-fa18-4bb1-a981-c5c359833031"). InnerVolumeSpecName "kube-api-access-cjrpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.633627 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1324b91-71af-44bd-9658-7727c10193c6-kube-api-access-lvtbs" (OuterVolumeSpecName: "kube-api-access-lvtbs") pod "d1324b91-71af-44bd-9658-7727c10193c6" (UID: "d1324b91-71af-44bd-9658-7727c10193c6"). InnerVolumeSpecName "kube-api-access-lvtbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.633891 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4757cf-18ca-406d-92ae-d975f16cb209-kube-api-access-bnhfb" (OuterVolumeSpecName: "kube-api-access-bnhfb") pod "5f4757cf-18ca-406d-92ae-d975f16cb209" (UID: "5f4757cf-18ca-406d-92ae-d975f16cb209"). InnerVolumeSpecName "kube-api-access-bnhfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.649133 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe692be-892a-46b0-b168-fa66ffce6762-kube-api-access-f9t9d" (OuterVolumeSpecName: "kube-api-access-f9t9d") pod "9fe692be-892a-46b0-b168-fa66ffce6762" (UID: "9fe692be-892a-46b0-b168-fa66ffce6762"). InnerVolumeSpecName "kube-api-access-f9t9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.727621 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f2262a-8a40-47de-9e14-d29ead210c3e-operator-scripts\") pod \"a3f2262a-8a40-47de-9e14-d29ead210c3e\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.727952 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knvbx\" (UniqueName: \"kubernetes.io/projected/a3f2262a-8a40-47de-9e14-d29ead210c3e-kube-api-access-knvbx\") pod \"a3f2262a-8a40-47de-9e14-d29ead210c3e\" (UID: \"a3f2262a-8a40-47de-9e14-d29ead210c3e\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.731815 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3f2262a-8a40-47de-9e14-d29ead210c3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3f2262a-8a40-47de-9e14-d29ead210c3e" (UID: "a3f2262a-8a40-47de-9e14-d29ead210c3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.732066 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-16d1-account-create-update-krscn" event={"ID":"5f4757cf-18ca-406d-92ae-d975f16cb209","Type":"ContainerDied","Data":"8ec768370a3a4e51083cfff7a57dcea48fba4eb0a12c5320ce02bc4bfab4bd72"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.732139 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ec768370a3a4e51083cfff7a57dcea48fba4eb0a12c5320ce02bc4bfab4bd72" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.732207 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-16d1-account-create-update-krscn" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.735568 4817 generic.go:334] "Generic (PLEG): container finished" podID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerID="57478e58c4952c9a190ce4ef2fc26d7345be350a1e244deee775c1d100af9f6a" exitCode=143 Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.735616 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1","Type":"ContainerDied","Data":"57478e58c4952c9a190ce4ef2fc26d7345be350a1e244deee775c1d100af9f6a"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.748044 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9t9d\" (UniqueName: \"kubernetes.io/projected/9fe692be-892a-46b0-b168-fa66ffce6762-kube-api-access-f9t9d\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751169 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvtbs\" (UniqueName: \"kubernetes.io/projected/d1324b91-71af-44bd-9658-7727c10193c6-kube-api-access-lvtbs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751457 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3f2262a-8a40-47de-9e14-d29ead210c3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751555 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjrpf\" (UniqueName: \"kubernetes.io/projected/74102797-fa18-4bb1-a981-c5c359833031-kube-api-access-cjrpf\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751565 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74102797-fa18-4bb1-a981-c5c359833031-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751576 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1324b91-71af-44bd-9658-7727c10193c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751586 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f4757cf-18ca-406d-92ae-d975f16cb209-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751596 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnhfb\" (UniqueName: \"kubernetes.io/projected/5f4757cf-18ca-406d-92ae-d975f16cb209-kube-api-access-bnhfb\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.751606 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fe692be-892a-46b0-b168-fa66ffce6762-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.763043 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f2262a-8a40-47de-9e14-d29ead210c3e-kube-api-access-knvbx" (OuterVolumeSpecName: "kube-api-access-knvbx") pod "a3f2262a-8a40-47de-9e14-d29ead210c3e" (UID: "a3f2262a-8a40-47de-9e14-d29ead210c3e"). InnerVolumeSpecName "kube-api-access-knvbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.772443 4817 generic.go:334] "Generic (PLEG): container finished" podID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerID="f4f22ca1b0a23418a0255001362927ff0e4b8f602cb4785a1133d170f894b648" exitCode=0 Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.775715 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62dsg" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.781290 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9f53839-25c4-4e8f-b9c7-1895ddff10d6","Type":"ContainerDied","Data":"f4f22ca1b0a23418a0255001362927ff0e4b8f602cb4785a1133d170f894b648"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.781333 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-62dsg" event={"ID":"74102797-fa18-4bb1-a981-c5c359833031","Type":"ContainerDied","Data":"7d74feff7b9b2c3bdc2dc65b1c9f287557bcca04f5abf6ee60041bad74949c01"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.781347 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d74feff7b9b2c3bdc2dc65b1c9f287557bcca04f5abf6ee60041bad74949c01" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.795906 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-063a-account-create-update-95fpn" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.795903 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-063a-account-create-update-95fpn" event={"ID":"9fe692be-892a-46b0-b168-fa66ffce6762","Type":"ContainerDied","Data":"c5f7dc477d522788edfbc80719cd75b2967b3358284e67de36c8dbcddf4bfb53"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.796080 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5f7dc477d522788edfbc80719cd75b2967b3358284e67de36c8dbcddf4bfb53" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.812255 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n67gd" event={"ID":"a3f2262a-8a40-47de-9e14-d29ead210c3e","Type":"ContainerDied","Data":"9511b9f8560cd698602c6682b1df9ac92dcb592ac1715f4c7a4548f78df1c1c9"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.812310 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9511b9f8560cd698602c6682b1df9ac92dcb592ac1715f4c7a4548f78df1c1c9" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.814422 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n67gd" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.824615 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vlrzf" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.824657 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vlrzf" event={"ID":"d1324b91-71af-44bd-9658-7727c10193c6","Type":"ContainerDied","Data":"945f3ac0f567a0db646f762c9b69485a02b31553b637377ffb61a85090c1801d"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.824692 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="945f3ac0f567a0db646f762c9b69485a02b31553b637377ffb61a85090c1801d" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.829380 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" event={"ID":"6afd8c02-96a9-44c5-a773-a137ab3df6e9","Type":"ContainerDied","Data":"84dbf4efaf3022e2d61704adfe7f47a267fd224bf2786da9dfaaa0ad980e7182"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.829415 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84dbf4efaf3022e2d61704adfe7f47a267fd224bf2786da9dfaaa0ad980e7182" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.829473 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5f5-account-create-update-2hjmb" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.834909 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerStarted","Data":"153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522"} Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.849692 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.853607 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knvbx\" (UniqueName: \"kubernetes.io/projected/a3f2262a-8a40-47de-9e14-d29ead210c3e-kube-api-access-knvbx\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.954560 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.954648 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfzpm\" (UniqueName: \"kubernetes.io/projected/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-kube-api-access-vfzpm\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.954744 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-logs\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.954828 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-config-data\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.954858 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-httpd-run\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.954934 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-scripts\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.955001 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-public-tls-certs\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.955026 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-combined-ca-bundle\") pod \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\" (UID: \"f9f53839-25c4-4e8f-b9c7-1895ddff10d6\") " Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.956083 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.957037 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-logs" (OuterVolumeSpecName: "logs") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.958497 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.958650 4817 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.959231 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.959267 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-kube-api-access-vfzpm" (OuterVolumeSpecName: "kube-api-access-vfzpm") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "kube-api-access-vfzpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.960018 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-scripts" (OuterVolumeSpecName: "scripts") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:11 crc kubenswrapper[4817]: I1128 14:48:11.996644 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.013764 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.019599 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-config-data" (OuterVolumeSpecName: "config-data") pod "f9f53839-25c4-4e8f-b9c7-1895ddff10d6" (UID: "f9f53839-25c4-4e8f-b9c7-1895ddff10d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.059948 4817 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.060167 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.060298 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.060380 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfzpm\" (UniqueName: \"kubernetes.io/projected/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-kube-api-access-vfzpm\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.060464 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.060580 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f53839-25c4-4e8f-b9c7-1895ddff10d6-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.089060 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.151158 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-c9859cc8b-qb8tq"] Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.154881 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6afd8c02-96a9-44c5-a773-a137ab3df6e9" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.154955 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6afd8c02-96a9-44c5-a773-a137ab3df6e9" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155055 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f2262a-8a40-47de-9e14-d29ead210c3e" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155107 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f2262a-8a40-47de-9e14-d29ead210c3e" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155174 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-httpd" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155221 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-httpd" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155270 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1324b91-71af-44bd-9658-7727c10193c6" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155316 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1324b91-71af-44bd-9658-7727c10193c6" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155368 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-log" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155413 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-log" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155471 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74102797-fa18-4bb1-a981-c5c359833031" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155522 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="74102797-fa18-4bb1-a981-c5c359833031" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155586 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4757cf-18ca-406d-92ae-d975f16cb209" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155633 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4757cf-18ca-406d-92ae-d975f16cb209" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: E1128 14:48:12.155701 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe692be-892a-46b0-b168-fa66ffce6762" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.155764 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe692be-892a-46b0-b168-fa66ffce6762" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156113 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f2262a-8a40-47de-9e14-d29ead210c3e" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156209 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4757cf-18ca-406d-92ae-d975f16cb209" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156277 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-httpd" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156326 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6afd8c02-96a9-44c5-a773-a137ab3df6e9" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156385 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fe692be-892a-46b0-b168-fa66ffce6762" containerName="mariadb-account-create-update" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156447 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" containerName="glance-log" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156538 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1324b91-71af-44bd-9658-7727c10193c6" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.156603 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="74102797-fa18-4bb1-a981-c5c359833031" containerName="mariadb-database-create" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.157231 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.162388 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.167438 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.168915 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-jmpmj" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.174634 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.196447 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-c9859cc8b-qb8tq"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.264028 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-combined-ca-bundle\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.264100 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flmhs\" (UniqueName: \"kubernetes.io/projected/801e2540-f948-4568-8437-6c5268542400-kube-api-access-flmhs\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.264156 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.264295 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data-custom\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.311585 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5bfc884db-x8mp7"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.312984 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.322466 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.326797 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-cxp27"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.328550 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.341308 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5bfc884db-x8mp7"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.364041 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-cxp27"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375128 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data-custom\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375213 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-config\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375295 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-combined-ca-bundle\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375341 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcfk5\" (UniqueName: \"kubernetes.io/projected/610bb394-f0ff-4876-bf71-86aacfa8f690-kube-api-access-wcfk5\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375431 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375474 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375535 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-combined-ca-bundle\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375556 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flmhs\" (UniqueName: \"kubernetes.io/projected/801e2540-f948-4568-8437-6c5268542400-kube-api-access-flmhs\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375574 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375592 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9svvw\" (UniqueName: \"kubernetes.io/projected/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-kube-api-access-9svvw\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375624 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375653 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375675 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.375743 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data-custom\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.381829 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-combined-ca-bundle\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.384994 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data-custom\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.386158 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.404528 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flmhs\" (UniqueName: \"kubernetes.io/projected/801e2540-f948-4568-8437-6c5268542400-kube-api-access-flmhs\") pod \"heat-engine-c9859cc8b-qb8tq\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.417923 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-54c7d598db-xws2q"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.419575 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.421149 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.451816 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-54c7d598db-xws2q"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.478165 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.479313 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-combined-ca-bundle\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.479446 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-config\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.479697 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.479830 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-combined-ca-bundle\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.479916 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcfk5\" (UniqueName: \"kubernetes.io/projected/610bb394-f0ff-4876-bf71-86aacfa8f690-kube-api-access-wcfk5\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480026 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480117 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480214 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480291 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9svvw\" (UniqueName: \"kubernetes.io/projected/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-kube-api-access-9svvw\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480363 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480425 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480523 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data-custom\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480597 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data-custom\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480682 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s47q\" (UniqueName: \"kubernetes.io/projected/80f4f160-440c-4978-a14c-13e952700dce-kube-api-access-8s47q\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480963 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.481754 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.480313 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-config\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.482502 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.482609 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.501467 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcfk5\" (UniqueName: \"kubernetes.io/projected/610bb394-f0ff-4876-bf71-86aacfa8f690-kube-api-access-wcfk5\") pod \"dnsmasq-dns-7756b9d78c-cxp27\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.507442 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-combined-ca-bundle\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.508381 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9svvw\" (UniqueName: \"kubernetes.io/projected/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-kube-api-access-9svvw\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.511297 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data-custom\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.526273 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data\") pod \"heat-cfnapi-5bfc884db-x8mp7\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.564268 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.583997 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s47q\" (UniqueName: \"kubernetes.io/projected/80f4f160-440c-4978-a14c-13e952700dce-kube-api-access-8s47q\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.584039 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-combined-ca-bundle\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.584096 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.584211 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data-custom\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.594950 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data-custom\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.596150 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-combined-ca-bundle\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.600684 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.612872 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s47q\" (UniqueName: \"kubernetes.io/projected/80f4f160-440c-4978-a14c-13e952700dce-kube-api-access-8s47q\") pod \"heat-api-54c7d598db-xws2q\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.766916 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.877785 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerStarted","Data":"a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031"} Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.882471 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f9f53839-25c4-4e8f-b9c7-1895ddff10d6","Type":"ContainerDied","Data":"ce198414b18af0d6b69e8ba0585f8f86425dec46b7ebbdf3fab5e0aa387bdc35"} Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.882559 4817 scope.go:117] "RemoveContainer" containerID="f4f22ca1b0a23418a0255001362927ff0e4b8f602cb4785a1133d170f894b648" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.882857 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.892171 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.960928 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.967940 4817 scope.go:117] "RemoveContainer" containerID="6fa052b7b40112b5990ebbff5d9a4b45c427e0cab7a5e5ecd75d6779e0bb83da" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.978319 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.986781 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.988593 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.996653 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 14:48:12 crc kubenswrapper[4817]: I1128 14:48:12.997683 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.000377 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.092236 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4f615c7-7cc5-4339-8bdc-e2974582e564-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.092311 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmmp9\" (UniqueName: \"kubernetes.io/projected/c4f615c7-7cc5-4339-8bdc-e2974582e564-kube-api-access-zmmp9\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.092377 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4f615c7-7cc5-4339-8bdc-e2974582e564-logs\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.093185 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-c9859cc8b-qb8tq"] Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.094868 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-scripts\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.094904 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-config-data\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.095018 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.095176 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.095245 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: W1128 14:48:13.174142 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod610bb394_f0ff_4876_bf71_86aacfa8f690.slice/crio-037990dc8156ba9da45f18cb7e61b330a964a5f2c9d815684ce22b0a9b209295 WatchSource:0}: Error finding container 037990dc8156ba9da45f18cb7e61b330a964a5f2c9d815684ce22b0a9b209295: Status 404 returned error can't find the container with id 037990dc8156ba9da45f18cb7e61b330a964a5f2c9d815684ce22b0a9b209295 Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.199732 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.199840 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4f615c7-7cc5-4339-8bdc-e2974582e564-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.199897 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmmp9\" (UniqueName: \"kubernetes.io/projected/c4f615c7-7cc5-4339-8bdc-e2974582e564-kube-api-access-zmmp9\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.199936 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4f615c7-7cc5-4339-8bdc-e2974582e564-logs\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.199967 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-scripts\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.199987 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-config-data\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.200023 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.200093 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.200705 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.202572 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c4f615c7-7cc5-4339-8bdc-e2974582e564-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.204531 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4f615c7-7cc5-4339-8bdc-e2974582e564-logs\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.214658 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-config-data\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.223694 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-scripts\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.227433 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmmp9\" (UniqueName: \"kubernetes.io/projected/c4f615c7-7cc5-4339-8bdc-e2974582e564-kube-api-access-zmmp9\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.233340 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.234786 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4f615c7-7cc5-4339-8bdc-e2974582e564-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.260206 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-cxp27"] Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.283227 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c4f615c7-7cc5-4339-8bdc-e2974582e564\") " pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.324121 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.372224 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5bfc884db-x8mp7"] Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.503632 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-54c7d598db-xws2q"] Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.748606 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9f53839-25c4-4e8f-b9c7-1895ddff10d6" path="/var/lib/kubelet/pods/f9f53839-25c4-4e8f-b9c7-1895ddff10d6/volumes" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.895057 4817 generic.go:334] "Generic (PLEG): container finished" podID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerID="a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7" exitCode=0 Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.895459 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" event={"ID":"610bb394-f0ff-4876-bf71-86aacfa8f690","Type":"ContainerDied","Data":"a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7"} Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.895534 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" event={"ID":"610bb394-f0ff-4876-bf71-86aacfa8f690","Type":"ContainerStarted","Data":"037990dc8156ba9da45f18cb7e61b330a964a5f2c9d815684ce22b0a9b209295"} Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.898952 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-54c7d598db-xws2q" event={"ID":"80f4f160-440c-4978-a14c-13e952700dce","Type":"ContainerStarted","Data":"b37b0d1413741809b3c18e9a1f60453d94b8a996b6b97ccb7f0d538cf94805be"} Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.900353 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" event={"ID":"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932","Type":"ContainerStarted","Data":"a5eca88111315d9243d09f5d068f00b7a8c534eb704f3ad9bc95f8fa80ad8f8d"} Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.902540 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c9859cc8b-qb8tq" event={"ID":"801e2540-f948-4568-8437-6c5268542400","Type":"ContainerStarted","Data":"6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3"} Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.902603 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c9859cc8b-qb8tq" event={"ID":"801e2540-f948-4568-8437-6c5268542400","Type":"ContainerStarted","Data":"c350fb209571c297eb15747e120589201f96dff58a293764002f7ce08a05a0a0"} Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.903465 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.941395 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-c9859cc8b-qb8tq" podStartSLOduration=1.941371074 podStartE2EDuration="1.941371074s" podCreationTimestamp="2025-11-28 14:48:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:13.937360495 +0000 UTC m=+1116.525338761" watchObservedRunningTime="2025-11-28 14:48:13.941371074 +0000 UTC m=+1116.529349340" Nov 28 14:48:13 crc kubenswrapper[4817]: I1128 14:48:13.964510 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 14:48:14 crc kubenswrapper[4817]: I1128 14:48:14.919576 4817 generic.go:334] "Generic (PLEG): container finished" podID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerID="7d6ec80e36c3086c48828d66ba9659a186575524e70c5be9355705832797e3d6" exitCode=0 Nov 28 14:48:14 crc kubenswrapper[4817]: I1128 14:48:14.920092 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1","Type":"ContainerDied","Data":"7d6ec80e36c3086c48828d66ba9659a186575524e70c5be9355705832797e3d6"} Nov 28 14:48:14 crc kubenswrapper[4817]: I1128 14:48:14.921255 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c4f615c7-7cc5-4339-8bdc-e2974582e564","Type":"ContainerStarted","Data":"57a549e7e41bca3f998f512984f97b1de3676f1b71f1c0522f303a5cbd818399"} Nov 28 14:48:14 crc kubenswrapper[4817]: I1128 14:48:14.954061 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" event={"ID":"610bb394-f0ff-4876-bf71-86aacfa8f690","Type":"ContainerStarted","Data":"35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739"} Nov 28 14:48:14 crc kubenswrapper[4817]: I1128 14:48:14.954495 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:14 crc kubenswrapper[4817]: I1128 14:48:14.973080 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" podStartSLOduration=2.973065107 podStartE2EDuration="2.973065107s" podCreationTimestamp="2025-11-28 14:48:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:14.971592091 +0000 UTC m=+1117.559570387" watchObservedRunningTime="2025-11-28 14:48:14.973065107 +0000 UTC m=+1117.561043373" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.155087 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.270818 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-internal-tls-certs\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271263 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-logs\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271317 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-scripts\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271350 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-config-data\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271445 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-combined-ca-bundle\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271476 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271522 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-httpd-run\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.271555 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgtd4\" (UniqueName: \"kubernetes.io/projected/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-kube-api-access-kgtd4\") pod \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\" (UID: \"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1\") " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.272102 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-logs" (OuterVolumeSpecName: "logs") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.272292 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.273317 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.280749 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-scripts" (OuterVolumeSpecName: "scripts") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.298958 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.299058 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-kube-api-access-kgtd4" (OuterVolumeSpecName: "kube-api-access-kgtd4") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "kube-api-access-kgtd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.330245 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.374285 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.374915 4817 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.374938 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgtd4\" (UniqueName: \"kubernetes.io/projected/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-kube-api-access-kgtd4\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.374951 4817 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.374962 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.374975 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.375000 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.402386 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-config-data" (OuterVolumeSpecName: "config-data") pod "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" (UID: "356ce6f1-31c2-4016-b6cd-a6737ddfb1a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.422332 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.477274 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.477316 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.811634 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-68nks"] Nov 28 14:48:15 crc kubenswrapper[4817]: E1128 14:48:15.812770 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-log" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.812791 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-log" Nov 28 14:48:15 crc kubenswrapper[4817]: E1128 14:48:15.812828 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-httpd" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.812837 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-httpd" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.813219 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-httpd" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.813254 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" containerName="glance-log" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.814561 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.817191 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.817462 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.817677 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-h88jc" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.835757 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-68nks"] Nov 28 14:48:15 crc kubenswrapper[4817]: E1128 14:48:15.873071 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod356ce6f1_31c2_4016_b6cd_a6737ddfb1a1.slice/crio-af321c9de28d682572ce61939bc206feb56547cfcc0fb2b598c534185238b284\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod356ce6f1_31c2_4016_b6cd_a6737ddfb1a1.slice\": RecentStats: unable to find data in memory cache]" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.992362 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"356ce6f1-31c2-4016-b6cd-a6737ddfb1a1","Type":"ContainerDied","Data":"af321c9de28d682572ce61939bc206feb56547cfcc0fb2b598c534185238b284"} Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.992755 4817 scope.go:117] "RemoveContainer" containerID="7d6ec80e36c3086c48828d66ba9659a186575524e70c5be9355705832797e3d6" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.992880 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.994893 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-config-data\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.994951 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6fcg\" (UniqueName: \"kubernetes.io/projected/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-kube-api-access-m6fcg\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.995006 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-scripts\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:15 crc kubenswrapper[4817]: I1128 14:48:15.995052 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.000503 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c4f615c7-7cc5-4339-8bdc-e2974582e564","Type":"ContainerStarted","Data":"b25c1a682d07bc765ec87e7326268879a2d3598625e050c075f9159c98ab991a"} Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.000541 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c4f615c7-7cc5-4339-8bdc-e2974582e564","Type":"ContainerStarted","Data":"25ee1392e0da14d0c71c6b702d82e369accca23911a635ce8524decfe3c446e8"} Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.005501 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerStarted","Data":"a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8"} Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.005763 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-central-agent" containerID="cri-o://de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948" gracePeriod=30 Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.005802 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="proxy-httpd" containerID="cri-o://a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8" gracePeriod=30 Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.005816 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="sg-core" containerID="cri-o://a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031" gracePeriod=30 Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.005845 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-notification-agent" containerID="cri-o://153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522" gracePeriod=30 Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.028783 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.039369 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.055543 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.057457 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.064426 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.064780 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.072954 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.077137 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.077117958 podStartE2EDuration="4.077117958s" podCreationTimestamp="2025-11-28 14:48:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:16.075339154 +0000 UTC m=+1118.663317420" watchObservedRunningTime="2025-11-28 14:48:16.077117958 +0000 UTC m=+1118.665096234" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.104853 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.105231 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-config-data\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.105299 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6fcg\" (UniqueName: \"kubernetes.io/projected/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-kube-api-access-m6fcg\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.105415 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-scripts\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.115477 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.116219 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-config-data\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.117816 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-scripts\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.133109 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.616655027 podStartE2EDuration="8.133090509s" podCreationTimestamp="2025-11-28 14:48:08 +0000 UTC" firstStartedPulling="2025-11-28 14:48:09.647660596 +0000 UTC m=+1112.235638862" lastFinishedPulling="2025-11-28 14:48:14.164096078 +0000 UTC m=+1116.752074344" observedRunningTime="2025-11-28 14:48:16.102744555 +0000 UTC m=+1118.690722821" watchObservedRunningTime="2025-11-28 14:48:16.133090509 +0000 UTC m=+1118.721068775" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.160934 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6fcg\" (UniqueName: \"kubernetes.io/projected/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-kube-api-access-m6fcg\") pod \"nova-cell0-conductor-db-sync-68nks\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.168202 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206479 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206550 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206648 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0451aa8-c32f-46e7-a9d8-574cb7005894-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206675 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfrwk\" (UniqueName: \"kubernetes.io/projected/d0451aa8-c32f-46e7-a9d8-574cb7005894-kube-api-access-lfrwk\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206699 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0451aa8-c32f-46e7-a9d8-574cb7005894-logs\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206755 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206801 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.206833 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308071 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308125 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308218 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308281 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308337 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0451aa8-c32f-46e7-a9d8-574cb7005894-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308357 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfrwk\" (UniqueName: \"kubernetes.io/projected/d0451aa8-c32f-46e7-a9d8-574cb7005894-kube-api-access-lfrwk\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308376 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0451aa8-c32f-46e7-a9d8-574cb7005894-logs\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.308396 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.309121 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.309384 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0451aa8-c32f-46e7-a9d8-574cb7005894-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.309707 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0451aa8-c32f-46e7-a9d8-574cb7005894-logs\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.312200 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.313637 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.314433 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.317851 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0451aa8-c32f-46e7-a9d8-574cb7005894-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.329934 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfrwk\" (UniqueName: \"kubernetes.io/projected/d0451aa8-c32f-46e7-a9d8-574cb7005894-kube-api-access-lfrwk\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.365034 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d0451aa8-c32f-46e7-a9d8-574cb7005894\") " pod="openstack/glance-default-internal-api-0" Nov 28 14:48:16 crc kubenswrapper[4817]: I1128 14:48:16.382699 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.016928 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb11a253-9431-42b9-8d11-ad971f486253" containerID="a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8" exitCode=0 Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.017262 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb11a253-9431-42b9-8d11-ad971f486253" containerID="a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031" exitCode=2 Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.017003 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerDied","Data":"a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8"} Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.017305 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerDied","Data":"a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031"} Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.017318 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerDied","Data":"153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522"} Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.017276 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb11a253-9431-42b9-8d11-ad971f486253" containerID="153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522" exitCode=0 Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.272539 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.293477 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85df6b4fc9-7c5nt" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.385705 4817 scope.go:117] "RemoveContainer" containerID="57478e58c4952c9a190ce4ef2fc26d7345be350a1e244deee775c1d100af9f6a" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.755921 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.757400 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="356ce6f1-31c2-4016-b6cd-a6737ddfb1a1" path="/var/lib/kubelet/pods/356ce6f1-31c2-4016-b6cd-a6737ddfb1a1/volumes" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.842598 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-run-httpd\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.842960 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-combined-ca-bundle\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.843039 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2gh8\" (UniqueName: \"kubernetes.io/projected/eb11a253-9431-42b9-8d11-ad971f486253-kube-api-access-b2gh8\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.843134 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-config-data\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.843198 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.843233 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-sg-core-conf-yaml\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.843258 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-scripts\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.843301 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-log-httpd\") pod \"eb11a253-9431-42b9-8d11-ad971f486253\" (UID: \"eb11a253-9431-42b9-8d11-ad971f486253\") " Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.844107 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.844520 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.844546 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb11a253-9431-42b9-8d11-ad971f486253-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.851387 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb11a253-9431-42b9-8d11-ad971f486253-kube-api-access-b2gh8" (OuterVolumeSpecName: "kube-api-access-b2gh8") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "kube-api-access-b2gh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.853089 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-scripts" (OuterVolumeSpecName: "scripts") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.887039 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.939428 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.946851 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.946882 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2gh8\" (UniqueName: \"kubernetes.io/projected/eb11a253-9431-42b9-8d11-ad971f486253-kube-api-access-b2gh8\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.946894 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.946903 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:17 crc kubenswrapper[4817]: I1128 14:48:17.953601 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-config-data" (OuterVolumeSpecName: "config-data") pod "eb11a253-9431-42b9-8d11-ad971f486253" (UID: "eb11a253-9431-42b9-8d11-ad971f486253"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.037951 4817 generic.go:334] "Generic (PLEG): container finished" podID="eb11a253-9431-42b9-8d11-ad971f486253" containerID="de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948" exitCode=0 Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.038022 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerDied","Data":"de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948"} Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.038050 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb11a253-9431-42b9-8d11-ad971f486253","Type":"ContainerDied","Data":"f7664cce90990e38d44a9be2ebf22249ef872088adac1d55790bfd15ddfb3439"} Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.038069 4817 scope.go:117] "RemoveContainer" containerID="a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.038165 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.040451 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-68nks"] Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.048540 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11a253-9431-42b9-8d11-ad971f486253-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.049082 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-54c7d598db-xws2q" event={"ID":"80f4f160-440c-4978-a14c-13e952700dce","Type":"ContainerStarted","Data":"c9b575518aa7a07348e079ceaf4228ed3450bf9a892055f33db0b1bc4c709a21"} Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.050203 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.068407 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" event={"ID":"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932","Type":"ContainerStarted","Data":"4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088"} Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.068551 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.071563 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-54c7d598db-xws2q" podStartSLOduration=2.209422005 podStartE2EDuration="6.07154846s" podCreationTimestamp="2025-11-28 14:48:12 +0000 UTC" firstStartedPulling="2025-11-28 14:48:13.526463156 +0000 UTC m=+1116.114441422" lastFinishedPulling="2025-11-28 14:48:17.388589611 +0000 UTC m=+1119.976567877" observedRunningTime="2025-11-28 14:48:18.07075445 +0000 UTC m=+1120.658732706" watchObservedRunningTime="2025-11-28 14:48:18.07154846 +0000 UTC m=+1120.659526726" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.103615 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" podStartSLOduration=2.096360425 podStartE2EDuration="6.103592956s" podCreationTimestamp="2025-11-28 14:48:12 +0000 UTC" firstStartedPulling="2025-11-28 14:48:13.396322332 +0000 UTC m=+1115.984300598" lastFinishedPulling="2025-11-28 14:48:17.403554863 +0000 UTC m=+1119.991533129" observedRunningTime="2025-11-28 14:48:18.091938126 +0000 UTC m=+1120.679916402" watchObservedRunningTime="2025-11-28 14:48:18.103592956 +0000 UTC m=+1120.691571222" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.116812 4817 scope.go:117] "RemoveContainer" containerID="a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.122347 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.156219 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.160242 4817 scope.go:117] "RemoveContainer" containerID="153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179107 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.179489 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="proxy-httpd" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179505 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="proxy-httpd" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.179520 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="sg-core" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179528 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="sg-core" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.179544 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-central-agent" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179550 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-central-agent" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.179567 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-notification-agent" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179573 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-notification-agent" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179771 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-central-agent" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179790 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="ceilometer-notification-agent" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179799 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="proxy-httpd" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.179808 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb11a253-9431-42b9-8d11-ad971f486253" containerName="sg-core" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.181338 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.187470 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.187654 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.193365 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.201292 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.242403 4817 scope.go:117] "RemoveContainer" containerID="de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.252552 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-run-httpd\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.252620 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-log-httpd\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.252933 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.253005 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.253037 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-config-data\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.253158 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzjg6\" (UniqueName: \"kubernetes.io/projected/4823eb3a-b603-40ed-8832-ba2838a18679-kube-api-access-fzjg6\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.253219 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-scripts\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.270747 4817 scope.go:117] "RemoveContainer" containerID="a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.272638 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8\": container with ID starting with a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8 not found: ID does not exist" containerID="a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.272711 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8"} err="failed to get container status \"a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8\": rpc error: code = NotFound desc = could not find container \"a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8\": container with ID starting with a347618cb10d620ec7de77e1a832088a7d852c71c253ca8af0cdd594d50770f8 not found: ID does not exist" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.272765 4817 scope.go:117] "RemoveContainer" containerID="a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.273134 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031\": container with ID starting with a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031 not found: ID does not exist" containerID="a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.273179 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031"} err="failed to get container status \"a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031\": rpc error: code = NotFound desc = could not find container \"a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031\": container with ID starting with a7e4f249c8e389217018c773b4f9ae96616bef90f6c6aba55b2e0d08baa60031 not found: ID does not exist" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.273197 4817 scope.go:117] "RemoveContainer" containerID="153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.273636 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522\": container with ID starting with 153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522 not found: ID does not exist" containerID="153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.273665 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522"} err="failed to get container status \"153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522\": rpc error: code = NotFound desc = could not find container \"153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522\": container with ID starting with 153607e167b7edb6db1a90355e46a9b3d4a48fc0cba5990030cf857543d3e522 not found: ID does not exist" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.273681 4817 scope.go:117] "RemoveContainer" containerID="de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948" Nov 28 14:48:18 crc kubenswrapper[4817]: E1128 14:48:18.273890 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948\": container with ID starting with de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948 not found: ID does not exist" containerID="de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.273935 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948"} err="failed to get container status \"de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948\": rpc error: code = NotFound desc = could not find container \"de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948\": container with ID starting with de13d51ce47e9ecc82130941bffdaa1164d6d550260b5de53e46492a8c904948 not found: ID does not exist" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.356480 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzjg6\" (UniqueName: \"kubernetes.io/projected/4823eb3a-b603-40ed-8832-ba2838a18679-kube-api-access-fzjg6\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.356865 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-scripts\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.356891 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-run-httpd\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.356945 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-log-httpd\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.357049 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.357096 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.357121 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-config-data\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.357669 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-log-httpd\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.357760 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-run-httpd\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.362761 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-config-data\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.363654 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.369292 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-scripts\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.375489 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.378757 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzjg6\" (UniqueName: \"kubernetes.io/projected/4823eb3a-b603-40ed-8832-ba2838a18679-kube-api-access-fzjg6\") pod \"ceilometer-0\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " pod="openstack/ceilometer-0" Nov 28 14:48:18 crc kubenswrapper[4817]: I1128 14:48:18.522662 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.072206 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:19 crc kubenswrapper[4817]: W1128 14:48:19.073843 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4823eb3a_b603_40ed_8832_ba2838a18679.slice/crio-69bc08addf80f7b0df0138a4530df6c0974f54192e262b90d756cfdaa9c54bd5 WatchSource:0}: Error finding container 69bc08addf80f7b0df0138a4530df6c0974f54192e262b90d756cfdaa9c54bd5: Status 404 returned error can't find the container with id 69bc08addf80f7b0df0138a4530df6c0974f54192e262b90d756cfdaa9c54bd5 Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.095281 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerStarted","Data":"69bc08addf80f7b0df0138a4530df6c0974f54192e262b90d756cfdaa9c54bd5"} Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.096502 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-68nks" event={"ID":"f7cf2d7d-acea-4cc4-bff9-a216dd46be04","Type":"ContainerStarted","Data":"5e602bc9268fb792a1b6bc36ceaa88938d64b4927faf9bdec3269d7fdd376fe4"} Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.098066 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d0451aa8-c32f-46e7-a9d8-574cb7005894","Type":"ContainerStarted","Data":"69b6685257c6cbf04f272c4505c08185e0be4be7235dd1a2b859a75b29b7a8a1"} Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.098109 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d0451aa8-c32f-46e7-a9d8-574cb7005894","Type":"ContainerStarted","Data":"8b94629a9d34ae92350e79fb3f4fd0eab3c409acc417f381d54cd19379e407ac"} Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.302402 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7585599b46-rxrsf"] Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.303688 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.331760 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5b47c555cb-jwdgc"] Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.332909 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.375117 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-config-data-custom\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.375220 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-combined-ca-bundle\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.375333 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgdk8\" (UniqueName: \"kubernetes.io/projected/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-kube-api-access-lgdk8\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.375396 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-config-data\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.382089 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7585599b46-rxrsf"] Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.416389 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-cbb8658c5-w9djk"] Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.417479 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.443303 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5b47c555cb-jwdgc"] Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.467791 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cbb8658c5-w9djk"] Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479451 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgdk8\" (UniqueName: \"kubernetes.io/projected/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-kube-api-access-lgdk8\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479556 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-config-data\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479629 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-combined-ca-bundle\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479658 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data-custom\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479689 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9xpt\" (UniqueName: \"kubernetes.io/projected/de7d0394-4866-4e91-a58d-d6852c5a5166-kube-api-access-x9xpt\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479768 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-config-data-custom\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479829 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-combined-ca-bundle\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.479861 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.498985 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-config-data\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.506562 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-config-data-custom\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.507128 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgdk8\" (UniqueName: \"kubernetes.io/projected/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-kube-api-access-lgdk8\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.510421 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde8e8ec-30e7-4913-83bb-2babaaf11f9a-combined-ca-bundle\") pod \"heat-engine-7585599b46-rxrsf\" (UID: \"bde8e8ec-30e7-4913-83bb-2babaaf11f9a\") " pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582063 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data-custom\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582119 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582153 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-combined-ca-bundle\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582177 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np2pl\" (UniqueName: \"kubernetes.io/projected/b292215e-e03a-4000-bc8a-3e8f68c9ba54-kube-api-access-np2pl\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582202 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data-custom\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582224 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9xpt\" (UniqueName: \"kubernetes.io/projected/de7d0394-4866-4e91-a58d-d6852c5a5166-kube-api-access-x9xpt\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582273 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-combined-ca-bundle\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.582323 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.589690 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data-custom\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.591301 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.601871 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-combined-ca-bundle\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.609987 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9xpt\" (UniqueName: \"kubernetes.io/projected/de7d0394-4866-4e91-a58d-d6852c5a5166-kube-api-access-x9xpt\") pod \"heat-api-5b47c555cb-jwdgc\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.684000 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.684255 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np2pl\" (UniqueName: \"kubernetes.io/projected/b292215e-e03a-4000-bc8a-3e8f68c9ba54-kube-api-access-np2pl\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.684287 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-combined-ca-bundle\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.684460 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data-custom\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.688461 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.688803 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data-custom\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.690173 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-combined-ca-bundle\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.705440 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np2pl\" (UniqueName: \"kubernetes.io/projected/b292215e-e03a-4000-bc8a-3e8f68c9ba54-kube-api-access-np2pl\") pod \"heat-cfnapi-cbb8658c5-w9djk\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.728113 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.744391 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.754185 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb11a253-9431-42b9-8d11-ad971f486253" path="/var/lib/kubelet/pods/eb11a253-9431-42b9-8d11-ad971f486253/volumes" Nov 28 14:48:19 crc kubenswrapper[4817]: I1128 14:48:19.764390 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:20 crc kubenswrapper[4817]: I1128 14:48:20.112020 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d0451aa8-c32f-46e7-a9d8-574cb7005894","Type":"ContainerStarted","Data":"664c8ba8f019a81fec6fdc8c6e1bc773e803ad01e0d1b3663a57fd095cde8e8d"} Nov 28 14:48:20 crc kubenswrapper[4817]: I1128 14:48:20.117365 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerStarted","Data":"7d8ff591e4bc0608f41d84d054f03f7473ba0a735813602a885c54089f34899f"} Nov 28 14:48:20 crc kubenswrapper[4817]: I1128 14:48:20.143135 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.143114009 podStartE2EDuration="4.143114009s" podCreationTimestamp="2025-11-28 14:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:20.12909241 +0000 UTC m=+1122.717070696" watchObservedRunningTime="2025-11-28 14:48:20.143114009 +0000 UTC m=+1122.731092275" Nov 28 14:48:20 crc kubenswrapper[4817]: I1128 14:48:20.336494 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7585599b46-rxrsf"] Nov 28 14:48:20 crc kubenswrapper[4817]: I1128 14:48:20.471860 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cbb8658c5-w9djk"] Nov 28 14:48:20 crc kubenswrapper[4817]: I1128 14:48:20.483864 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5b47c555cb-jwdgc"] Nov 28 14:48:20 crc kubenswrapper[4817]: W1128 14:48:20.507388 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb292215e_e03a_4000_bc8a_3e8f68c9ba54.slice/crio-b6e97d86d5c05a9f0328c032524b56f16099900b1dbcd920001c43d605f588df WatchSource:0}: Error finding container b6e97d86d5c05a9f0328c032524b56f16099900b1dbcd920001c43d605f588df: Status 404 returned error can't find the container with id b6e97d86d5c05a9f0328c032524b56f16099900b1dbcd920001c43d605f588df Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.135492 4817 generic.go:334] "Generic (PLEG): container finished" podID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerID="2593fdb1c7fed275a3b9ac71e0eaeaeb68d0f0f8ac7824db64328b960f5c2101" exitCode=1 Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.135549 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b47c555cb-jwdgc" event={"ID":"de7d0394-4866-4e91-a58d-d6852c5a5166","Type":"ContainerDied","Data":"2593fdb1c7fed275a3b9ac71e0eaeaeb68d0f0f8ac7824db64328b960f5c2101"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.135898 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b47c555cb-jwdgc" event={"ID":"de7d0394-4866-4e91-a58d-d6852c5a5166","Type":"ContainerStarted","Data":"26994ffefb4cdcff0ea616ed2ffd3f9fb4b20ff025b9702e2e04ce919c98c289"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.136074 4817 scope.go:117] "RemoveContainer" containerID="2593fdb1c7fed275a3b9ac71e0eaeaeb68d0f0f8ac7824db64328b960f5c2101" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.149235 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7585599b46-rxrsf" event={"ID":"bde8e8ec-30e7-4913-83bb-2babaaf11f9a","Type":"ContainerStarted","Data":"df12f3111b9defaa6a7675d0e3cee617481448d88496b162ae7dda16e197dcd7"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.149282 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7585599b46-rxrsf" event={"ID":"bde8e8ec-30e7-4913-83bb-2babaaf11f9a","Type":"ContainerStarted","Data":"673d1d0fc35604dcb68e41d80221782ff4e442c48093521337bf05175f189322"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.149823 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.153569 4817 generic.go:334] "Generic (PLEG): container finished" podID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerID="f1b75cee7273c85f26489b8cefca279aceda310d7e3bdbfa96869945b60b4695" exitCode=1 Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.154402 4817 scope.go:117] "RemoveContainer" containerID="f1b75cee7273c85f26489b8cefca279aceda310d7e3bdbfa96869945b60b4695" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.153640 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" event={"ID":"b292215e-e03a-4000-bc8a-3e8f68c9ba54","Type":"ContainerDied","Data":"f1b75cee7273c85f26489b8cefca279aceda310d7e3bdbfa96869945b60b4695"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.156148 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" event={"ID":"b292215e-e03a-4000-bc8a-3e8f68c9ba54","Type":"ContainerStarted","Data":"b6e97d86d5c05a9f0328c032524b56f16099900b1dbcd920001c43d605f588df"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.157861 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerStarted","Data":"c8c5453bd20a9510e727839689ba809fa64e59692197e54aa7cc3750ac229c69"} Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.191787 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7585599b46-rxrsf" podStartSLOduration=2.191760972 podStartE2EDuration="2.191760972s" podCreationTimestamp="2025-11-28 14:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:21.179014896 +0000 UTC m=+1123.766993172" watchObservedRunningTime="2025-11-28 14:48:21.191760972 +0000 UTC m=+1123.779739238" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.453313 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.530292 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5bfc884db-x8mp7"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.530497 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" podUID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" containerName="heat-cfnapi" containerID="cri-o://4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088" gracePeriod=60 Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.548436 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-54c7d598db-xws2q"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.548644 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-54c7d598db-xws2q" podUID="80f4f160-440c-4978-a14c-13e952700dce" containerName="heat-api" containerID="cri-o://c9b575518aa7a07348e079ceaf4228ed3450bf9a892055f33db0b1bc4c709a21" gracePeriod=60 Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.608196 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5696cfb998-r25ks"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.618534 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.624701 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.624802 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.636372 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5696cfb998-r25ks"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.714832 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6f9fcf49b7-fdlpq"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.716805 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.731228 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.731455 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.733264 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-internal-tls-certs\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.733317 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-config-data\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.733361 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-public-tls-certs\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.733444 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2sdc\" (UniqueName: \"kubernetes.io/projected/11d0830f-fa4e-4204-b32d-3ef9553325f1-kube-api-access-c2sdc\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.733468 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-config-data-custom\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.733486 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-combined-ca-bundle\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.832652 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6f9fcf49b7-fdlpq"] Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835380 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-combined-ca-bundle\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835444 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-config-data\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835473 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-public-tls-certs\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835498 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-config-data-custom\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835525 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2sdc\" (UniqueName: \"kubernetes.io/projected/11d0830f-fa4e-4204-b32d-3ef9553325f1-kube-api-access-c2sdc\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835550 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-config-data-custom\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835565 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-combined-ca-bundle\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835614 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-internal-tls-certs\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835640 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-internal-tls-certs\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835668 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-config-data\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.835691 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6v6g\" (UniqueName: \"kubernetes.io/projected/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-kube-api-access-b6v6g\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.836371 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-public-tls-certs\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.857657 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-internal-tls-certs\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.860614 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2sdc\" (UniqueName: \"kubernetes.io/projected/11d0830f-fa4e-4204-b32d-3ef9553325f1-kube-api-access-c2sdc\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.876939 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-config-data-custom\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.877900 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-public-tls-certs\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.878449 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-combined-ca-bundle\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.894510 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11d0830f-fa4e-4204-b32d-3ef9553325f1-config-data\") pod \"heat-cfnapi-5696cfb998-r25ks\" (UID: \"11d0830f-fa4e-4204-b32d-3ef9553325f1\") " pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938471 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-internal-tls-certs\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938542 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6v6g\" (UniqueName: \"kubernetes.io/projected/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-kube-api-access-b6v6g\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938594 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-combined-ca-bundle\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938621 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-config-data\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938645 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-public-tls-certs\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938668 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-config-data-custom\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.938972 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.946658 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-config-data-custom\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.947043 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-combined-ca-bundle\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.947161 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-public-tls-certs\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.971534 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6v6g\" (UniqueName: \"kubernetes.io/projected/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-kube-api-access-b6v6g\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.971813 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-config-data\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:21 crc kubenswrapper[4817]: I1128 14:48:21.981617 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b12bb76-1c8c-4ea3-afd6-b1ff967ae174-internal-tls-certs\") pod \"heat-api-6f9fcf49b7-fdlpq\" (UID: \"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174\") " pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.138390 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.179073 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" event={"ID":"b292215e-e03a-4000-bc8a-3e8f68c9ba54","Type":"ContainerStarted","Data":"5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb"} Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.179858 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.188978 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b47c555cb-jwdgc" event={"ID":"de7d0394-4866-4e91-a58d-d6852c5a5166","Type":"ContainerStarted","Data":"5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37"} Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.189755 4817 status_manager.go:317] "Container readiness changed for unknown container" pod="openstack/heat-api-5b47c555cb-jwdgc" containerID="cri-o://2593fdb1c7fed275a3b9ac71e0eaeaeb68d0f0f8ac7824db64328b960f5c2101" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.189769 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.211380 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" podStartSLOduration=3.211359575 podStartE2EDuration="3.211359575s" podCreationTimestamp="2025-11-28 14:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:22.20350912 +0000 UTC m=+1124.791487386" watchObservedRunningTime="2025-11-28 14:48:22.211359575 +0000 UTC m=+1124.799337841" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.234117 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5b47c555cb-jwdgc" podStartSLOduration=3.234078369 podStartE2EDuration="3.234078369s" podCreationTimestamp="2025-11-28 14:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:22.222967173 +0000 UTC m=+1124.810945429" watchObservedRunningTime="2025-11-28 14:48:22.234078369 +0000 UTC m=+1124.822056635" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.471551 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5696cfb998-r25ks"] Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.567887 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.661958 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-8cwtv"] Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.662214 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerName="dnsmasq-dns" containerID="cri-o://12296b6bbd7a948c76c48f76ea624ecf6b21180469d50c877b35fa3771591782" gracePeriod=10 Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.691577 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6f9fcf49b7-fdlpq"] Nov 28 14:48:22 crc kubenswrapper[4817]: W1128 14:48:22.786883 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b12bb76_1c8c_4ea3_afd6_b1ff967ae174.slice/crio-9d352d254bcfdd92c66117bce2b4a536356fe21152c73f8d3dd79d0b8517bd9e WatchSource:0}: Error finding container 9d352d254bcfdd92c66117bce2b4a536356fe21152c73f8d3dd79d0b8517bd9e: Status 404 returned error can't find the container with id 9d352d254bcfdd92c66117bce2b4a536356fe21152c73f8d3dd79d0b8517bd9e Nov 28 14:48:22 crc kubenswrapper[4817]: I1128 14:48:22.895039 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-54c7d598db-xws2q" podUID="80f4f160-440c-4978-a14c-13e952700dce" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.181:8004/healthcheck\": dial tcp 10.217.0.181:8004: connect: connection refused" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.072832 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.168376 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data-custom\") pod \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.168449 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9svvw\" (UniqueName: \"kubernetes.io/projected/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-kube-api-access-9svvw\") pod \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.168465 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data\") pod \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.168505 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-combined-ca-bundle\") pod \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\" (UID: \"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.175917 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-kube-api-access-9svvw" (OuterVolumeSpecName: "kube-api-access-9svvw") pod "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" (UID: "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932"). InnerVolumeSpecName "kube-api-access-9svvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.176505 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" (UID: "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.207888 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" (UID: "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.208295 4817 generic.go:334] "Generic (PLEG): container finished" podID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" containerID="4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088" exitCode=0 Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.208351 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.208362 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" event={"ID":"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932","Type":"ContainerDied","Data":"4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.208389 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" event={"ID":"dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932","Type":"ContainerDied","Data":"a5eca88111315d9243d09f5d068f00b7a8c534eb704f3ad9bc95f8fa80ad8f8d"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.208405 4817 scope.go:117] "RemoveContainer" containerID="4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.212554 4817 generic.go:334] "Generic (PLEG): container finished" podID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerID="12296b6bbd7a948c76c48f76ea624ecf6b21180469d50c877b35fa3771591782" exitCode=0 Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.212632 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" event={"ID":"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42","Type":"ContainerDied","Data":"12296b6bbd7a948c76c48f76ea624ecf6b21180469d50c877b35fa3771591782"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.226284 4817 generic.go:334] "Generic (PLEG): container finished" podID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerID="5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb" exitCode=1 Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.226377 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" event={"ID":"b292215e-e03a-4000-bc8a-3e8f68c9ba54","Type":"ContainerDied","Data":"5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.227084 4817 scope.go:117] "RemoveContainer" containerID="5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb" Nov 28 14:48:23 crc kubenswrapper[4817]: E1128 14:48:23.227484 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cbb8658c5-w9djk_openstack(b292215e-e03a-4000-bc8a-3e8f68c9ba54)\"" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.230558 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f9fcf49b7-fdlpq" event={"ID":"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174","Type":"ContainerStarted","Data":"9d352d254bcfdd92c66117bce2b4a536356fe21152c73f8d3dd79d0b8517bd9e"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.235970 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5696cfb998-r25ks" event={"ID":"11d0830f-fa4e-4204-b32d-3ef9553325f1","Type":"ContainerStarted","Data":"bf362ae316743c3823b1e3d5a04f397c127198859c1185d15f12fbcd6935a392"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.239263 4817 scope.go:117] "RemoveContainer" containerID="4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088" Nov 28 14:48:23 crc kubenswrapper[4817]: E1128 14:48:23.239821 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088\": container with ID starting with 4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088 not found: ID does not exist" containerID="4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.239862 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088"} err="failed to get container status \"4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088\": rpc error: code = NotFound desc = could not find container \"4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088\": container with ID starting with 4fa84acce11ed7c83bef4bcfdb90809a9d7675b67ac4f387d97ec4c7a2971088 not found: ID does not exist" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.239887 4817 scope.go:117] "RemoveContainer" containerID="f1b75cee7273c85f26489b8cefca279aceda310d7e3bdbfa96869945b60b4695" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.243056 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerStarted","Data":"d4f875185ce32a382be0723d101164b3b49c2689b45c19311c49049c32834702"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.249026 4817 generic.go:334] "Generic (PLEG): container finished" podID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerID="5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37" exitCode=1 Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.249130 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b47c555cb-jwdgc" event={"ID":"de7d0394-4866-4e91-a58d-d6852c5a5166","Type":"ContainerDied","Data":"5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.251021 4817 scope.go:117] "RemoveContainer" containerID="5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37" Nov 28 14:48:23 crc kubenswrapper[4817]: E1128 14:48:23.251323 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5b47c555cb-jwdgc_openstack(de7d0394-4866-4e91-a58d-d6852c5a5166)\"" pod="openstack/heat-api-5b47c555cb-jwdgc" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.254842 4817 generic.go:334] "Generic (PLEG): container finished" podID="80f4f160-440c-4978-a14c-13e952700dce" containerID="c9b575518aa7a07348e079ceaf4228ed3450bf9a892055f33db0b1bc4c709a21" exitCode=0 Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.254898 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-54c7d598db-xws2q" event={"ID":"80f4f160-440c-4978-a14c-13e952700dce","Type":"ContainerDied","Data":"c9b575518aa7a07348e079ceaf4228ed3450bf9a892055f33db0b1bc4c709a21"} Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.261213 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data" (OuterVolumeSpecName: "config-data") pod "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" (UID: "dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.271219 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.271244 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9svvw\" (UniqueName: \"kubernetes.io/projected/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-kube-api-access-9svvw\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.271255 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.271264 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.300715 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.326961 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.327024 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.409445 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.411448 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.411631 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.480052 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-sb\") pod \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.480133 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-svc\") pod \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.480195 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-swift-storage-0\") pod \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.480287 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-config\") pod \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.480392 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9wv7\" (UniqueName: \"kubernetes.io/projected/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-kube-api-access-p9wv7\") pod \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.480437 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-nb\") pod \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\" (UID: \"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.486853 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-kube-api-access-p9wv7" (OuterVolumeSpecName: "kube-api-access-p9wv7") pod "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" (UID: "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42"). InnerVolumeSpecName "kube-api-access-p9wv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.527124 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" (UID: "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.540244 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-config" (OuterVolumeSpecName: "config") pod "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" (UID: "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.570618 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" (UID: "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.574462 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" (UID: "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.575506 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" (UID: "58c5c818-fcb3-4bf2-b8ca-f8a880d82f42"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.581918 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data-custom\") pod \"80f4f160-440c-4978-a14c-13e952700dce\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.581999 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s47q\" (UniqueName: \"kubernetes.io/projected/80f4f160-440c-4978-a14c-13e952700dce-kube-api-access-8s47q\") pod \"80f4f160-440c-4978-a14c-13e952700dce\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582101 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data\") pod \"80f4f160-440c-4978-a14c-13e952700dce\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582172 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-combined-ca-bundle\") pod \"80f4f160-440c-4978-a14c-13e952700dce\" (UID: \"80f4f160-440c-4978-a14c-13e952700dce\") " Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582606 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582628 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582637 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582648 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582657 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9wv7\" (UniqueName: \"kubernetes.io/projected/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-kube-api-access-p9wv7\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.582668 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.586005 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "80f4f160-440c-4978-a14c-13e952700dce" (UID: "80f4f160-440c-4978-a14c-13e952700dce"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.587921 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80f4f160-440c-4978-a14c-13e952700dce-kube-api-access-8s47q" (OuterVolumeSpecName: "kube-api-access-8s47q") pod "80f4f160-440c-4978-a14c-13e952700dce" (UID: "80f4f160-440c-4978-a14c-13e952700dce"). InnerVolumeSpecName "kube-api-access-8s47q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.611790 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80f4f160-440c-4978-a14c-13e952700dce" (UID: "80f4f160-440c-4978-a14c-13e952700dce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.639867 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data" (OuterVolumeSpecName: "config-data") pod "80f4f160-440c-4978-a14c-13e952700dce" (UID: "80f4f160-440c-4978-a14c-13e952700dce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.684803 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.684839 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.684850 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s47q\" (UniqueName: \"kubernetes.io/projected/80f4f160-440c-4978-a14c-13e952700dce-kube-api-access-8s47q\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.684863 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f4f160-440c-4978-a14c-13e952700dce-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.769742 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5bfc884db-x8mp7"] Nov 28 14:48:23 crc kubenswrapper[4817]: I1128 14:48:23.769774 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5bfc884db-x8mp7"] Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.273494 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" event={"ID":"58c5c818-fcb3-4bf2-b8ca-f8a880d82f42","Type":"ContainerDied","Data":"dba1f683c2eab8e31f755cc2ac2cead7f2e085a553b99ea9d8ff10ebaa20f9f2"} Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.273544 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-8cwtv" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.276980 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f9fcf49b7-fdlpq" event={"ID":"2b12bb76-1c8c-4ea3-afd6-b1ff967ae174","Type":"ContainerStarted","Data":"fd0d9512b27b9158435c5926db049837e11ac088a847f93be24bf26a0f71d4a2"} Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.279292 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5696cfb998-r25ks" event={"ID":"11d0830f-fa4e-4204-b32d-3ef9553325f1","Type":"ContainerStarted","Data":"599c3065979b30194ee706b614dcc2e6fe767895ef62beb93ccec09d6f30c9d3"} Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.282024 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-54c7d598db-xws2q" event={"ID":"80f4f160-440c-4978-a14c-13e952700dce","Type":"ContainerDied","Data":"b37b0d1413741809b3c18e9a1f60453d94b8a996b6b97ccb7f0d538cf94805be"} Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.282193 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-54c7d598db-xws2q" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.282573 4817 scope.go:117] "RemoveContainer" containerID="5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb" Nov 28 14:48:24 crc kubenswrapper[4817]: E1128 14:48:24.282862 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cbb8658c5-w9djk_openstack(b292215e-e03a-4000-bc8a-3e8f68c9ba54)\"" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.282871 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.283614 4817 scope.go:117] "RemoveContainer" containerID="5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37" Nov 28 14:48:24 crc kubenswrapper[4817]: E1128 14:48:24.283818 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5b47c555cb-jwdgc_openstack(de7d0394-4866-4e91-a58d-d6852c5a5166)\"" pod="openstack/heat-api-5b47c555cb-jwdgc" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.283849 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.302661 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-8cwtv"] Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.320752 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-8cwtv"] Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.367075 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-54c7d598db-xws2q"] Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.379148 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-54c7d598db-xws2q"] Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.380967 4817 scope.go:117] "RemoveContainer" containerID="2593fdb1c7fed275a3b9ac71e0eaeaeb68d0f0f8ac7824db64328b960f5c2101" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.470500 4817 scope.go:117] "RemoveContainer" containerID="12296b6bbd7a948c76c48f76ea624ecf6b21180469d50c877b35fa3771591782" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.489591 4817 scope.go:117] "RemoveContainer" containerID="2d8b0fd8a128f6adbee545e6ae5a4facd16d48d151450327e0dc6ea471e99ab2" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.515670 4817 scope.go:117] "RemoveContainer" containerID="c9b575518aa7a07348e079ceaf4228ed3450bf9a892055f33db0b1bc4c709a21" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.745736 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.745795 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:24 crc kubenswrapper[4817]: I1128 14:48:24.765356 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.299642 4817 scope.go:117] "RemoveContainer" containerID="5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb" Nov 28 14:48:25 crc kubenswrapper[4817]: E1128 14:48:25.300054 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-cbb8658c5-w9djk_openstack(b292215e-e03a-4000-bc8a-3e8f68c9ba54)\"" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.309552 4817 scope.go:117] "RemoveContainer" containerID="5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37" Nov 28 14:48:25 crc kubenswrapper[4817]: E1128 14:48:25.309833 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5b47c555cb-jwdgc_openstack(de7d0394-4866-4e91-a58d-d6852c5a5166)\"" pod="openstack/heat-api-5b47c555cb-jwdgc" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.349766 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5696cfb998-r25ks" podStartSLOduration=4.349749309 podStartE2EDuration="4.349749309s" podCreationTimestamp="2025-11-28 14:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:25.342894479 +0000 UTC m=+1127.930872745" watchObservedRunningTime="2025-11-28 14:48:25.349749309 +0000 UTC m=+1127.937727575" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.361472 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6f9fcf49b7-fdlpq" podStartSLOduration=4.36143983 podStartE2EDuration="4.36143983s" podCreationTimestamp="2025-11-28 14:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:48:25.360902416 +0000 UTC m=+1127.948880682" watchObservedRunningTime="2025-11-28 14:48:25.36143983 +0000 UTC m=+1127.949418096" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.748806 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" path="/var/lib/kubelet/pods/58c5c818-fcb3-4bf2-b8ca-f8a880d82f42/volumes" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.749638 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80f4f160-440c-4978-a14c-13e952700dce" path="/var/lib/kubelet/pods/80f4f160-440c-4978-a14c-13e952700dce/volumes" Nov 28 14:48:25 crc kubenswrapper[4817]: I1128 14:48:25.750147 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" path="/var/lib/kubelet/pods/dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932/volumes" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.327468 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.327501 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.384161 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.384213 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.437491 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.439264 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:26 crc kubenswrapper[4817]: I1128 14:48:26.940569 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:27 crc kubenswrapper[4817]: I1128 14:48:27.139918 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:27 crc kubenswrapper[4817]: I1128 14:48:27.337353 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:27 crc kubenswrapper[4817]: I1128 14:48:27.337552 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:27 crc kubenswrapper[4817]: I1128 14:48:27.514931 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 14:48:27 crc kubenswrapper[4817]: I1128 14:48:27.515038 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:48:27 crc kubenswrapper[4817]: I1128 14:48:27.612321 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 14:48:29 crc kubenswrapper[4817]: I1128 14:48:29.388191 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:29 crc kubenswrapper[4817]: I1128 14:48:29.388565 4817 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 14:48:29 crc kubenswrapper[4817]: I1128 14:48:29.527674 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 14:48:32 crc kubenswrapper[4817]: I1128 14:48:32.515206 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:32 crc kubenswrapper[4817]: I1128 14:48:32.774062 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5bfc884db-x8mp7" podUID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.179:8000/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 14:48:33 crc kubenswrapper[4817]: I1128 14:48:33.977586 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6f9fcf49b7-fdlpq" Nov 28 14:48:33 crc kubenswrapper[4817]: I1128 14:48:33.983952 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5696cfb998-r25ks" Nov 28 14:48:34 crc kubenswrapper[4817]: I1128 14:48:34.059870 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5b47c555cb-jwdgc"] Nov 28 14:48:34 crc kubenswrapper[4817]: I1128 14:48:34.075887 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cbb8658c5-w9djk"] Nov 28 14:48:36 crc kubenswrapper[4817]: E1128 14:48:36.859174 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Nov 28 14:48:36 crc kubenswrapper[4817]: E1128 14:48:36.859646 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m6fcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-68nks_openstack(f7cf2d7d-acea-4cc4-bff9-a216dd46be04): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:48:36 crc kubenswrapper[4817]: E1128 14:48:36.861360 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-68nks" podUID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" Nov 28 14:48:36 crc kubenswrapper[4817]: I1128 14:48:36.979912 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:36 crc kubenswrapper[4817]: I1128 14:48:36.997748 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.067571 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data\") pod \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.067832 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data\") pod \"de7d0394-4866-4e91-a58d-d6852c5a5166\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.067931 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data-custom\") pod \"de7d0394-4866-4e91-a58d-d6852c5a5166\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.067996 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9xpt\" (UniqueName: \"kubernetes.io/projected/de7d0394-4866-4e91-a58d-d6852c5a5166-kube-api-access-x9xpt\") pod \"de7d0394-4866-4e91-a58d-d6852c5a5166\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.068152 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-combined-ca-bundle\") pod \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.068250 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np2pl\" (UniqueName: \"kubernetes.io/projected/b292215e-e03a-4000-bc8a-3e8f68c9ba54-kube-api-access-np2pl\") pod \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.068295 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data-custom\") pod \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\" (UID: \"b292215e-e03a-4000-bc8a-3e8f68c9ba54\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.068325 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-combined-ca-bundle\") pod \"de7d0394-4866-4e91-a58d-d6852c5a5166\" (UID: \"de7d0394-4866-4e91-a58d-d6852c5a5166\") " Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.077600 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b292215e-e03a-4000-bc8a-3e8f68c9ba54" (UID: "b292215e-e03a-4000-bc8a-3e8f68c9ba54"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.078071 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "de7d0394-4866-4e91-a58d-d6852c5a5166" (UID: "de7d0394-4866-4e91-a58d-d6852c5a5166"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.080080 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de7d0394-4866-4e91-a58d-d6852c5a5166-kube-api-access-x9xpt" (OuterVolumeSpecName: "kube-api-access-x9xpt") pod "de7d0394-4866-4e91-a58d-d6852c5a5166" (UID: "de7d0394-4866-4e91-a58d-d6852c5a5166"). InnerVolumeSpecName "kube-api-access-x9xpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.092007 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b292215e-e03a-4000-bc8a-3e8f68c9ba54-kube-api-access-np2pl" (OuterVolumeSpecName: "kube-api-access-np2pl") pod "b292215e-e03a-4000-bc8a-3e8f68c9ba54" (UID: "b292215e-e03a-4000-bc8a-3e8f68c9ba54"). InnerVolumeSpecName "kube-api-access-np2pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.103894 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b292215e-e03a-4000-bc8a-3e8f68c9ba54" (UID: "b292215e-e03a-4000-bc8a-3e8f68c9ba54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.117467 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de7d0394-4866-4e91-a58d-d6852c5a5166" (UID: "de7d0394-4866-4e91-a58d-d6852c5a5166"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.171672 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.171712 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np2pl\" (UniqueName: \"kubernetes.io/projected/b292215e-e03a-4000-bc8a-3e8f68c9ba54-kube-api-access-np2pl\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.171740 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.171753 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.171765 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.171778 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9xpt\" (UniqueName: \"kubernetes.io/projected/de7d0394-4866-4e91-a58d-d6852c5a5166-kube-api-access-x9xpt\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.173678 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data" (OuterVolumeSpecName: "config-data") pod "de7d0394-4866-4e91-a58d-d6852c5a5166" (UID: "de7d0394-4866-4e91-a58d-d6852c5a5166"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.193283 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data" (OuterVolumeSpecName: "config-data") pod "b292215e-e03a-4000-bc8a-3e8f68c9ba54" (UID: "b292215e-e03a-4000-bc8a-3e8f68c9ba54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.273058 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292215e-e03a-4000-bc8a-3e8f68c9ba54-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.273108 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de7d0394-4866-4e91-a58d-d6852c5a5166-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.457926 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5b47c555cb-jwdgc" event={"ID":"de7d0394-4866-4e91-a58d-d6852c5a5166","Type":"ContainerDied","Data":"26994ffefb4cdcff0ea616ed2ffd3f9fb4b20ff025b9702e2e04ce919c98c289"} Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.458002 4817 scope.go:117] "RemoveContainer" containerID="5f1f4decc1cdb6501c1ae805e8abbf682e034a8149fc1676e95d7766f38f3a37" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.457977 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5b47c555cb-jwdgc" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.459770 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.459787 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cbb8658c5-w9djk" event={"ID":"b292215e-e03a-4000-bc8a-3e8f68c9ba54","Type":"ContainerDied","Data":"b6e97d86d5c05a9f0328c032524b56f16099900b1dbcd920001c43d605f588df"} Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.471901 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-central-agent" containerID="cri-o://7d8ff591e4bc0608f41d84d054f03f7473ba0a735813602a885c54089f34899f" gracePeriod=30 Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.472409 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerStarted","Data":"ebd7349f99c50d54b3786e15ff777d98dcda28eaf8e64691f2f3c28940e247fb"} Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.472507 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="proxy-httpd" containerID="cri-o://ebd7349f99c50d54b3786e15ff777d98dcda28eaf8e64691f2f3c28940e247fb" gracePeriod=30 Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.472671 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-notification-agent" containerID="cri-o://c8c5453bd20a9510e727839689ba809fa64e59692197e54aa7cc3750ac229c69" gracePeriod=30 Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.472845 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.473747 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="sg-core" containerID="cri-o://d4f875185ce32a382be0723d101164b3b49c2689b45c19311c49049c32834702" gracePeriod=30 Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.489146 4817 scope.go:117] "RemoveContainer" containerID="5e535fd5d5e4552b32da9cebf63b6f326a3e47d9526d7eed1fff3cf80d9439fb" Nov 28 14:48:37 crc kubenswrapper[4817]: E1128 14:48:37.489468 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-68nks" podUID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.530193 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.351966054 podStartE2EDuration="19.530172366s" podCreationTimestamp="2025-11-28 14:48:18 +0000 UTC" firstStartedPulling="2025-11-28 14:48:19.0805892 +0000 UTC m=+1121.668567486" lastFinishedPulling="2025-11-28 14:48:33.258795532 +0000 UTC m=+1135.846773798" observedRunningTime="2025-11-28 14:48:37.526883414 +0000 UTC m=+1140.114861670" watchObservedRunningTime="2025-11-28 14:48:37.530172366 +0000 UTC m=+1140.118150632" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.569820 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5b47c555cb-jwdgc"] Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.589518 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5b47c555cb-jwdgc"] Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.600708 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cbb8658c5-w9djk"] Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.608424 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-cbb8658c5-w9djk"] Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.761097 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" path="/var/lib/kubelet/pods/b292215e-e03a-4000-bc8a-3e8f68c9ba54/volumes" Nov 28 14:48:37 crc kubenswrapper[4817]: I1128 14:48:37.761801 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" path="/var/lib/kubelet/pods/de7d0394-4866-4e91-a58d-d6852c5a5166/volumes" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.510733 4817 generic.go:334] "Generic (PLEG): container finished" podID="4823eb3a-b603-40ed-8832-ba2838a18679" containerID="ebd7349f99c50d54b3786e15ff777d98dcda28eaf8e64691f2f3c28940e247fb" exitCode=0 Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511064 4817 generic.go:334] "Generic (PLEG): container finished" podID="4823eb3a-b603-40ed-8832-ba2838a18679" containerID="d4f875185ce32a382be0723d101164b3b49c2689b45c19311c49049c32834702" exitCode=2 Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511078 4817 generic.go:334] "Generic (PLEG): container finished" podID="4823eb3a-b603-40ed-8832-ba2838a18679" containerID="c8c5453bd20a9510e727839689ba809fa64e59692197e54aa7cc3750ac229c69" exitCode=0 Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511087 4817 generic.go:334] "Generic (PLEG): container finished" podID="4823eb3a-b603-40ed-8832-ba2838a18679" containerID="7d8ff591e4bc0608f41d84d054f03f7473ba0a735813602a885c54089f34899f" exitCode=0 Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511184 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerDied","Data":"ebd7349f99c50d54b3786e15ff777d98dcda28eaf8e64691f2f3c28940e247fb"} Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511215 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerDied","Data":"d4f875185ce32a382be0723d101164b3b49c2689b45c19311c49049c32834702"} Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511228 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerDied","Data":"c8c5453bd20a9510e727839689ba809fa64e59692197e54aa7cc3750ac229c69"} Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.511241 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerDied","Data":"7d8ff591e4bc0608f41d84d054f03f7473ba0a735813602a885c54089f34899f"} Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.603052 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.701757 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-log-httpd\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.701950 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-sg-core-conf-yaml\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.701980 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-scripts\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.702049 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-combined-ca-bundle\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.702154 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-run-httpd\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.702186 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzjg6\" (UniqueName: \"kubernetes.io/projected/4823eb3a-b603-40ed-8832-ba2838a18679-kube-api-access-fzjg6\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.702239 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-config-data\") pod \"4823eb3a-b603-40ed-8832-ba2838a18679\" (UID: \"4823eb3a-b603-40ed-8832-ba2838a18679\") " Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.702522 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.703477 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.703555 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.709010 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-scripts" (OuterVolumeSpecName: "scripts") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.709440 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4823eb3a-b603-40ed-8832-ba2838a18679-kube-api-access-fzjg6" (OuterVolumeSpecName: "kube-api-access-fzjg6") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "kube-api-access-fzjg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.735021 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.807190 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.807223 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.807236 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4823eb3a-b603-40ed-8832-ba2838a18679-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.807476 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzjg6\" (UniqueName: \"kubernetes.io/projected/4823eb3a-b603-40ed-8832-ba2838a18679-kube-api-access-fzjg6\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.809007 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.829539 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-config-data" (OuterVolumeSpecName: "config-data") pod "4823eb3a-b603-40ed-8832-ba2838a18679" (UID: "4823eb3a-b603-40ed-8832-ba2838a18679"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.908632 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:38 crc kubenswrapper[4817]: I1128 14:48:38.908658 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4823eb3a-b603-40ed-8832-ba2838a18679-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.542007 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4823eb3a-b603-40ed-8832-ba2838a18679","Type":"ContainerDied","Data":"69bc08addf80f7b0df0138a4530df6c0974f54192e262b90d756cfdaa9c54bd5"} Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.542097 4817 scope.go:117] "RemoveContainer" containerID="ebd7349f99c50d54b3786e15ff777d98dcda28eaf8e64691f2f3c28940e247fb" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.542094 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.591914 4817 scope.go:117] "RemoveContainer" containerID="d4f875185ce32a382be0723d101164b3b49c2689b45c19311c49049c32834702" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.594040 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.603430 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.612830 4817 scope.go:117] "RemoveContainer" containerID="c8c5453bd20a9510e727839689ba809fa64e59692197e54aa7cc3750ac229c69" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.628071 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.628879 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.628904 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.628914 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="sg-core" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.628923 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="sg-core" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.628942 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="proxy-httpd" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.628951 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="proxy-httpd" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.628961 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-central-agent" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.628969 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-central-agent" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.628984 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-notification-agent" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.628992 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-notification-agent" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629017 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerName="dnsmasq-dns" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629023 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerName="dnsmasq-dns" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629034 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80f4f160-440c-4978-a14c-13e952700dce" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629039 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="80f4f160-440c-4978-a14c-13e952700dce" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629045 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629052 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629065 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629072 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629081 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629086 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629092 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerName="init" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629098 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerName="init" Nov 28 14:48:39 crc kubenswrapper[4817]: E1128 14:48:39.629108 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629114 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629339 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629359 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b292215e-e03a-4000-bc8a-3e8f68c9ba54" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629369 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="proxy-httpd" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629381 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="sg-core" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629390 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c5c818-fcb3-4bf2-b8ca-f8a880d82f42" containerName="dnsmasq-dns" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629398 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629407 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-notification-agent" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629424 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="80f4f160-440c-4978-a14c-13e952700dce" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629433 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" containerName="ceilometer-central-agent" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.629445 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd6ed3c8-ce3d-4ed1-ab4c-a899e7e69932" containerName="heat-cfnapi" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.630058 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="de7d0394-4866-4e91-a58d-d6852c5a5166" containerName="heat-api" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.632501 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.638415 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.638700 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.647097 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.655635 4817 scope.go:117] "RemoveContainer" containerID="7d8ff591e4bc0608f41d84d054f03f7473ba0a735813602a885c54089f34899f" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721129 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-config-data\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721174 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721208 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-run-httpd\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721465 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721505 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-scripts\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721733 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-log-httpd\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.721905 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4rq7\" (UniqueName: \"kubernetes.io/projected/b499e170-08ff-4285-96d9-a49da9ffa010-kube-api-access-t4rq7\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.749140 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4823eb3a-b603-40ed-8832-ba2838a18679" path="/var/lib/kubelet/pods/4823eb3a-b603-40ed-8832-ba2838a18679/volumes" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.762331 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7585599b46-rxrsf" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.804674 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-c9859cc8b-qb8tq"] Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.805036 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-c9859cc8b-qb8tq" podUID="801e2540-f948-4568-8437-6c5268542400" containerName="heat-engine" containerID="cri-o://6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" gracePeriod=60 Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.823554 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4rq7\" (UniqueName: \"kubernetes.io/projected/b499e170-08ff-4285-96d9-a49da9ffa010-kube-api-access-t4rq7\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.823900 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-config-data\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.823932 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.823960 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-run-httpd\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.824574 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-scripts\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.824615 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.824617 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-run-httpd\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.824804 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-log-httpd\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.825174 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-log-httpd\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.829360 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.829593 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.832800 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-scripts\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.837586 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-config-data\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.842978 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4rq7\" (UniqueName: \"kubernetes.io/projected/b499e170-08ff-4285-96d9-a49da9ffa010-kube-api-access-t4rq7\") pod \"ceilometer-0\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " pod="openstack/ceilometer-0" Nov 28 14:48:39 crc kubenswrapper[4817]: I1128 14:48:39.953393 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:48:40 crc kubenswrapper[4817]: I1128 14:48:40.428896 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:48:40 crc kubenswrapper[4817]: W1128 14:48:40.450460 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb499e170_08ff_4285_96d9_a49da9ffa010.slice/crio-59562377a462c7291e029745db73dd9e1903d20a2c0d36cd2b9e3b164fa0f0f0 WatchSource:0}: Error finding container 59562377a462c7291e029745db73dd9e1903d20a2c0d36cd2b9e3b164fa0f0f0: Status 404 returned error can't find the container with id 59562377a462c7291e029745db73dd9e1903d20a2c0d36cd2b9e3b164fa0f0f0 Nov 28 14:48:40 crc kubenswrapper[4817]: I1128 14:48:40.554403 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerStarted","Data":"59562377a462c7291e029745db73dd9e1903d20a2c0d36cd2b9e3b164fa0f0f0"} Nov 28 14:48:41 crc kubenswrapper[4817]: I1128 14:48:41.566047 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerStarted","Data":"ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6"} Nov 28 14:48:42 crc kubenswrapper[4817]: E1128 14:48:42.483123 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 28 14:48:42 crc kubenswrapper[4817]: E1128 14:48:42.506479 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 28 14:48:42 crc kubenswrapper[4817]: E1128 14:48:42.516768 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 28 14:48:42 crc kubenswrapper[4817]: E1128 14:48:42.516857 4817 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-c9859cc8b-qb8tq" podUID="801e2540-f948-4568-8437-6c5268542400" containerName="heat-engine" Nov 28 14:48:42 crc kubenswrapper[4817]: I1128 14:48:42.575575 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerStarted","Data":"a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5"} Nov 28 14:48:43 crc kubenswrapper[4817]: I1128 14:48:43.586279 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerStarted","Data":"c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975"} Nov 28 14:48:44 crc kubenswrapper[4817]: I1128 14:48:44.597702 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerStarted","Data":"bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387"} Nov 28 14:48:44 crc kubenswrapper[4817]: I1128 14:48:44.598042 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:48:44 crc kubenswrapper[4817]: I1128 14:48:44.630948 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9875061330000001 podStartE2EDuration="5.630923155s" podCreationTimestamp="2025-11-28 14:48:39 +0000 UTC" firstStartedPulling="2025-11-28 14:48:40.454614194 +0000 UTC m=+1143.042592460" lastFinishedPulling="2025-11-28 14:48:44.098031206 +0000 UTC m=+1146.686009482" observedRunningTime="2025-11-28 14:48:44.618427825 +0000 UTC m=+1147.206406091" watchObservedRunningTime="2025-11-28 14:48:44.630923155 +0000 UTC m=+1147.218901421" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.046948 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.141672 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data\") pod \"801e2540-f948-4568-8437-6c5268542400\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.142188 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-combined-ca-bundle\") pod \"801e2540-f948-4568-8437-6c5268542400\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.142247 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flmhs\" (UniqueName: \"kubernetes.io/projected/801e2540-f948-4568-8437-6c5268542400-kube-api-access-flmhs\") pod \"801e2540-f948-4568-8437-6c5268542400\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.142369 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data-custom\") pod \"801e2540-f948-4568-8437-6c5268542400\" (UID: \"801e2540-f948-4568-8437-6c5268542400\") " Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.147024 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "801e2540-f948-4568-8437-6c5268542400" (UID: "801e2540-f948-4568-8437-6c5268542400"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.150226 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/801e2540-f948-4568-8437-6c5268542400-kube-api-access-flmhs" (OuterVolumeSpecName: "kube-api-access-flmhs") pod "801e2540-f948-4568-8437-6c5268542400" (UID: "801e2540-f948-4568-8437-6c5268542400"). InnerVolumeSpecName "kube-api-access-flmhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.183206 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "801e2540-f948-4568-8437-6c5268542400" (UID: "801e2540-f948-4568-8437-6c5268542400"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.194457 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data" (OuterVolumeSpecName: "config-data") pod "801e2540-f948-4568-8437-6c5268542400" (UID: "801e2540-f948-4568-8437-6c5268542400"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.245123 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flmhs\" (UniqueName: \"kubernetes.io/projected/801e2540-f948-4568-8437-6c5268542400-kube-api-access-flmhs\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.245162 4817 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.245176 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.245188 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/801e2540-f948-4568-8437-6c5268542400-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.616676 4817 generic.go:334] "Generic (PLEG): container finished" podID="801e2540-f948-4568-8437-6c5268542400" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" exitCode=0 Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.616760 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c9859cc8b-qb8tq" event={"ID":"801e2540-f948-4568-8437-6c5268542400","Type":"ContainerDied","Data":"6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3"} Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.616795 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c9859cc8b-qb8tq" event={"ID":"801e2540-f948-4568-8437-6c5268542400","Type":"ContainerDied","Data":"c350fb209571c297eb15747e120589201f96dff58a293764002f7ce08a05a0a0"} Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.616819 4817 scope.go:117] "RemoveContainer" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.616817 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c9859cc8b-qb8tq" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.642782 4817 scope.go:117] "RemoveContainer" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" Nov 28 14:48:46 crc kubenswrapper[4817]: E1128 14:48:46.643269 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3\": container with ID starting with 6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3 not found: ID does not exist" containerID="6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.643327 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3"} err="failed to get container status \"6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3\": rpc error: code = NotFound desc = could not find container \"6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3\": container with ID starting with 6799d4c9f8144d766002301d3ae0142eaab4f59574e72e2bf2ccdcd582d84ee3 not found: ID does not exist" Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.666622 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-c9859cc8b-qb8tq"] Nov 28 14:48:46 crc kubenswrapper[4817]: I1128 14:48:46.677974 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-c9859cc8b-qb8tq"] Nov 28 14:48:47 crc kubenswrapper[4817]: I1128 14:48:47.747932 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="801e2540-f948-4568-8437-6c5268542400" path="/var/lib/kubelet/pods/801e2540-f948-4568-8437-6c5268542400/volumes" Nov 28 14:48:50 crc kubenswrapper[4817]: I1128 14:48:50.657782 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-68nks" event={"ID":"f7cf2d7d-acea-4cc4-bff9-a216dd46be04","Type":"ContainerStarted","Data":"cf73eb05681a447f3f54901052a036fb35dfc99ed41bb400ebf920a3dca94a8f"} Nov 28 14:48:50 crc kubenswrapper[4817]: I1128 14:48:50.688245 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-68nks" podStartSLOduration=3.381651674 podStartE2EDuration="35.68822689s" podCreationTimestamp="2025-11-28 14:48:15 +0000 UTC" firstStartedPulling="2025-11-28 14:48:18.054916886 +0000 UTC m=+1120.642895152" lastFinishedPulling="2025-11-28 14:48:50.361492102 +0000 UTC m=+1152.949470368" observedRunningTime="2025-11-28 14:48:50.67333566 +0000 UTC m=+1153.261313926" watchObservedRunningTime="2025-11-28 14:48:50.68822689 +0000 UTC m=+1153.276205146" Nov 28 14:49:01 crc kubenswrapper[4817]: I1128 14:49:01.753621 4817 generic.go:334] "Generic (PLEG): container finished" podID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" containerID="cf73eb05681a447f3f54901052a036fb35dfc99ed41bb400ebf920a3dca94a8f" exitCode=0 Nov 28 14:49:01 crc kubenswrapper[4817]: I1128 14:49:01.753709 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-68nks" event={"ID":"f7cf2d7d-acea-4cc4-bff9-a216dd46be04","Type":"ContainerDied","Data":"cf73eb05681a447f3f54901052a036fb35dfc99ed41bb400ebf920a3dca94a8f"} Nov 28 14:49:02 crc kubenswrapper[4817]: I1128 14:49:02.786318 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:02 crc kubenswrapper[4817]: I1128 14:49:02.787154 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-central-agent" containerID="cri-o://ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6" gracePeriod=30 Nov 28 14:49:02 crc kubenswrapper[4817]: I1128 14:49:02.787851 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-notification-agent" containerID="cri-o://a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5" gracePeriod=30 Nov 28 14:49:02 crc kubenswrapper[4817]: I1128 14:49:02.787875 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="sg-core" containerID="cri-o://c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975" gracePeriod=30 Nov 28 14:49:02 crc kubenswrapper[4817]: I1128 14:49:02.787845 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="proxy-httpd" containerID="cri-o://bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387" gracePeriod=30 Nov 28 14:49:02 crc kubenswrapper[4817]: I1128 14:49:02.814586 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.177219 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.276548 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-config-data\") pod \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.276605 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-combined-ca-bundle\") pod \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.276641 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-scripts\") pod \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.276734 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6fcg\" (UniqueName: \"kubernetes.io/projected/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-kube-api-access-m6fcg\") pod \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\" (UID: \"f7cf2d7d-acea-4cc4-bff9-a216dd46be04\") " Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.309263 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-kube-api-access-m6fcg" (OuterVolumeSpecName: "kube-api-access-m6fcg") pod "f7cf2d7d-acea-4cc4-bff9-a216dd46be04" (UID: "f7cf2d7d-acea-4cc4-bff9-a216dd46be04"). InnerVolumeSpecName "kube-api-access-m6fcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.309303 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-scripts" (OuterVolumeSpecName: "scripts") pod "f7cf2d7d-acea-4cc4-bff9-a216dd46be04" (UID: "f7cf2d7d-acea-4cc4-bff9-a216dd46be04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.328945 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-config-data" (OuterVolumeSpecName: "config-data") pod "f7cf2d7d-acea-4cc4-bff9-a216dd46be04" (UID: "f7cf2d7d-acea-4cc4-bff9-a216dd46be04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.348692 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7cf2d7d-acea-4cc4-bff9-a216dd46be04" (UID: "f7cf2d7d-acea-4cc4-bff9-a216dd46be04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.378607 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6fcg\" (UniqueName: \"kubernetes.io/projected/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-kube-api-access-m6fcg\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.378644 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.378655 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.378665 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7cf2d7d-acea-4cc4-bff9-a216dd46be04-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.771014 4817 generic.go:334] "Generic (PLEG): container finished" podID="b499e170-08ff-4285-96d9-a49da9ffa010" containerID="bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387" exitCode=0 Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.771048 4817 generic.go:334] "Generic (PLEG): container finished" podID="b499e170-08ff-4285-96d9-a49da9ffa010" containerID="c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975" exitCode=2 Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.771059 4817 generic.go:334] "Generic (PLEG): container finished" podID="b499e170-08ff-4285-96d9-a49da9ffa010" containerID="ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6" exitCode=0 Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.771093 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerDied","Data":"bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387"} Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.771140 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerDied","Data":"c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975"} Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.771156 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerDied","Data":"ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6"} Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.772868 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-68nks" event={"ID":"f7cf2d7d-acea-4cc4-bff9-a216dd46be04","Type":"ContainerDied","Data":"5e602bc9268fb792a1b6bc36ceaa88938d64b4927faf9bdec3269d7fdd376fe4"} Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.772898 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e602bc9268fb792a1b6bc36ceaa88938d64b4927faf9bdec3269d7fdd376fe4" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.773122 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-68nks" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.881395 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 14:49:03 crc kubenswrapper[4817]: E1128 14:49:03.881938 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" containerName="nova-cell0-conductor-db-sync" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.881958 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" containerName="nova-cell0-conductor-db-sync" Nov 28 14:49:03 crc kubenswrapper[4817]: E1128 14:49:03.881978 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="801e2540-f948-4568-8437-6c5268542400" containerName="heat-engine" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.881986 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="801e2540-f948-4568-8437-6c5268542400" containerName="heat-engine" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.882196 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="801e2540-f948-4568-8437-6c5268542400" containerName="heat-engine" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.882224 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" containerName="nova-cell0-conductor-db-sync" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.883424 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.886277 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.886521 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-h88jc" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.892330 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.989341 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dff079-658b-4cb4-aa29-42eb2c020705-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.989380 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dff079-658b-4cb4-aa29-42eb2c020705-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:03 crc kubenswrapper[4817]: I1128 14:49:03.989457 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b44p5\" (UniqueName: \"kubernetes.io/projected/98dff079-658b-4cb4-aa29-42eb2c020705-kube-api-access-b44p5\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.091346 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b44p5\" (UniqueName: \"kubernetes.io/projected/98dff079-658b-4cb4-aa29-42eb2c020705-kube-api-access-b44p5\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.091513 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dff079-658b-4cb4-aa29-42eb2c020705-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.091545 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dff079-658b-4cb4-aa29-42eb2c020705-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.104877 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dff079-658b-4cb4-aa29-42eb2c020705-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.110023 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b44p5\" (UniqueName: \"kubernetes.io/projected/98dff079-658b-4cb4-aa29-42eb2c020705-kube-api-access-b44p5\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.118520 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dff079-658b-4cb4-aa29-42eb2c020705-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"98dff079-658b-4cb4-aa29-42eb2c020705\") " pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.204591 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.365921 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499527 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-config-data\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499574 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-combined-ca-bundle\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499616 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-run-httpd\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499780 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-log-httpd\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499836 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-sg-core-conf-yaml\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499917 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4rq7\" (UniqueName: \"kubernetes.io/projected/b499e170-08ff-4285-96d9-a49da9ffa010-kube-api-access-t4rq7\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.499993 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-scripts\") pod \"b499e170-08ff-4285-96d9-a49da9ffa010\" (UID: \"b499e170-08ff-4285-96d9-a49da9ffa010\") " Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.500421 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.500591 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.500941 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.500963 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b499e170-08ff-4285-96d9-a49da9ffa010-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.505893 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-scripts" (OuterVolumeSpecName: "scripts") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.506451 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b499e170-08ff-4285-96d9-a49da9ffa010-kube-api-access-t4rq7" (OuterVolumeSpecName: "kube-api-access-t4rq7") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "kube-api-access-t4rq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.555759 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.603018 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4rq7\" (UniqueName: \"kubernetes.io/projected/b499e170-08ff-4285-96d9-a49da9ffa010-kube-api-access-t4rq7\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.603046 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.603057 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.620471 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-config-data" (OuterVolumeSpecName: "config-data") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.658908 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b499e170-08ff-4285-96d9-a49da9ffa010" (UID: "b499e170-08ff-4285-96d9-a49da9ffa010"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.704385 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.704415 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b499e170-08ff-4285-96d9-a49da9ffa010-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.720905 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 14:49:04 crc kubenswrapper[4817]: W1128 14:49:04.723780 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98dff079_658b_4cb4_aa29_42eb2c020705.slice/crio-2d9872dfcb6aaf9bfb30800577d6b434f1bb19bd493ffccf08a2c5be5079c569 WatchSource:0}: Error finding container 2d9872dfcb6aaf9bfb30800577d6b434f1bb19bd493ffccf08a2c5be5079c569: Status 404 returned error can't find the container with id 2d9872dfcb6aaf9bfb30800577d6b434f1bb19bd493ffccf08a2c5be5079c569 Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.781801 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"98dff079-658b-4cb4-aa29-42eb2c020705","Type":"ContainerStarted","Data":"2d9872dfcb6aaf9bfb30800577d6b434f1bb19bd493ffccf08a2c5be5079c569"} Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.788671 4817 generic.go:334] "Generic (PLEG): container finished" podID="b499e170-08ff-4285-96d9-a49da9ffa010" containerID="a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5" exitCode=0 Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.788709 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerDied","Data":"a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5"} Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.788764 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b499e170-08ff-4285-96d9-a49da9ffa010","Type":"ContainerDied","Data":"59562377a462c7291e029745db73dd9e1903d20a2c0d36cd2b9e3b164fa0f0f0"} Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.788784 4817 scope.go:117] "RemoveContainer" containerID="bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.788946 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.821026 4817 scope.go:117] "RemoveContainer" containerID="c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.825596 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.833457 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.848667 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.849050 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="proxy-httpd" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849066 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="proxy-httpd" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.849076 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-notification-agent" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849082 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-notification-agent" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.849115 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="sg-core" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849123 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="sg-core" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.849136 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-central-agent" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849143 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-central-agent" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849334 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-notification-agent" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849359 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="ceilometer-central-agent" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849367 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="sg-core" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.849378 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" containerName="proxy-httpd" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.851448 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.853670 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.853940 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.870802 4817 scope.go:117] "RemoveContainer" containerID="a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.893975 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.909759 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-scripts\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.909805 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.909847 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.909870 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-log-httpd\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.909938 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-config-data\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.909981 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-run-httpd\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.910011 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c7pt\" (UniqueName: \"kubernetes.io/projected/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-kube-api-access-6c7pt\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.910348 4817 scope.go:117] "RemoveContainer" containerID="ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.991367 4817 scope.go:117] "RemoveContainer" containerID="bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.992187 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387\": container with ID starting with bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387 not found: ID does not exist" containerID="bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.992244 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387"} err="failed to get container status \"bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387\": rpc error: code = NotFound desc = could not find container \"bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387\": container with ID starting with bd40685ac533be7bd21681e1888cb8e06c9ee9123b58f24f4a53a9c165776387 not found: ID does not exist" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.992299 4817 scope.go:117] "RemoveContainer" containerID="c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.992672 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975\": container with ID starting with c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975 not found: ID does not exist" containerID="c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.992707 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975"} err="failed to get container status \"c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975\": rpc error: code = NotFound desc = could not find container \"c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975\": container with ID starting with c730419908704647cf354273fb90e821bf347e6bb6283dd2ea8dd0a5b9b49975 not found: ID does not exist" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.992765 4817 scope.go:117] "RemoveContainer" containerID="a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.993038 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5\": container with ID starting with a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5 not found: ID does not exist" containerID="a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.993072 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5"} err="failed to get container status \"a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5\": rpc error: code = NotFound desc = could not find container \"a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5\": container with ID starting with a25c8252c89d4cc6516d48fc5f418b57760e2af08efbccaf81442ee87d7ccae5 not found: ID does not exist" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.993096 4817 scope.go:117] "RemoveContainer" containerID="ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6" Nov 28 14:49:04 crc kubenswrapper[4817]: E1128 14:49:04.993374 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6\": container with ID starting with ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6 not found: ID does not exist" containerID="ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6" Nov 28 14:49:04 crc kubenswrapper[4817]: I1128 14:49:04.993413 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6"} err="failed to get container status \"ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6\": rpc error: code = NotFound desc = could not find container \"ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6\": container with ID starting with ce0554bcee148c4aa8f4c35a2a0280307f4a2fded93571da392832c45e7f6aa6 not found: ID does not exist" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011567 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-config-data\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011654 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-run-httpd\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011705 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c7pt\" (UniqueName: \"kubernetes.io/projected/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-kube-api-access-6c7pt\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011799 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-scripts\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011829 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011873 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.011895 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-log-httpd\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.012260 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-run-httpd\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.012318 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-log-httpd\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.018410 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-scripts\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.019600 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-config-data\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.020471 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.021182 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.030253 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c7pt\" (UniqueName: \"kubernetes.io/projected/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-kube-api-access-6c7pt\") pod \"ceilometer-0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.168963 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.663170 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.756519 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b499e170-08ff-4285-96d9-a49da9ffa010" path="/var/lib/kubelet/pods/b499e170-08ff-4285-96d9-a49da9ffa010/volumes" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.800518 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"98dff079-658b-4cb4-aa29-42eb2c020705","Type":"ContainerStarted","Data":"94b1558d7b876433b30a764812f1709bdb419ca647a83a3b6b528b7f8751cb24"} Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.801117 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:05 crc kubenswrapper[4817]: I1128 14:49:05.804067 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerStarted","Data":"d3e84e0038ee2e7e50f3a02bf1183ebc5d60026a0eed418b78d33d4fbb8aa3f4"} Nov 28 14:49:06 crc kubenswrapper[4817]: I1128 14:49:06.814305 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerStarted","Data":"a176b6d644c4b1253ca8fd7b48b44f3f4bb2668a409ce0bd25f39f2212e196f1"} Nov 28 14:49:07 crc kubenswrapper[4817]: I1128 14:49:07.771518 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=4.771497181 podStartE2EDuration="4.771497181s" podCreationTimestamp="2025-11-28 14:49:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:05.831126 +0000 UTC m=+1168.419104276" watchObservedRunningTime="2025-11-28 14:49:07.771497181 +0000 UTC m=+1170.359475447" Nov 28 14:49:07 crc kubenswrapper[4817]: I1128 14:49:07.827283 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerStarted","Data":"33e8a4a7124d1c0e454b1c76980a764a4580f33aff0136fc6bbcf32a32042d66"} Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.237539 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.896062 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-f52xr"] Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.897618 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-f52xr"] Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.897715 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.901259 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.901540 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.910607 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerStarted","Data":"e565165da831a8f924eb15a51d3d281e0465304876aafe359039d345234bdcf9"} Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.940544 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.941968 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.959243 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.981071 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.982566 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:09 crc kubenswrapper[4817]: I1128 14:49:09.992120 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.012987 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041756 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-scripts\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041856 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-config-data\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041877 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq2m5\" (UniqueName: \"kubernetes.io/projected/c93d5fea-7f33-47a2-916a-80e9c68a89f2-kube-api-access-tq2m5\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041902 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwkh4\" (UniqueName: \"kubernetes.io/projected/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-kube-api-access-kwkh4\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041950 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041965 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.041993 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.071779 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.075526 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.080231 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.082032 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.105269 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.138855 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.140288 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143740 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-config-data\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143777 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-config-data\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143798 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143824 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143846 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-config-data\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143866 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq2m5\" (UniqueName: \"kubernetes.io/projected/c93d5fea-7f33-47a2-916a-80e9c68a89f2-kube-api-access-tq2m5\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143884 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-config-data\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143902 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwkh4\" (UniqueName: \"kubernetes.io/projected/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-kube-api-access-kwkh4\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143917 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnm8w\" (UniqueName: \"kubernetes.io/projected/3dc7b7d9-372e-4aa4-9776-f67e78497501-kube-api-access-rnm8w\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143960 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.143978 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144004 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144028 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144050 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9msj\" (UniqueName: \"kubernetes.io/projected/9e6712c0-048c-4432-aad1-2dc9e280ae0b-kube-api-access-p9msj\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144081 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-scripts\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144100 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwhnx\" (UniqueName: \"kubernetes.io/projected/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-kube-api-access-dwhnx\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144126 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-logs\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144340 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc7b7d9-372e-4aa4-9776-f67e78497501-logs\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.144619 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.153984 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-scripts\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.156064 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.163289 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.163429 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.163523 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-config-data\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.183903 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq2m5\" (UniqueName: \"kubernetes.io/projected/c93d5fea-7f33-47a2-916a-80e9c68a89f2-kube-api-access-tq2m5\") pod \"nova-cell0-cell-mapping-f52xr\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.195340 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwkh4\" (UniqueName: \"kubernetes.io/projected/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-kube-api-access-kwkh4\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.196565 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.217678 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lkgcd"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.219185 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.225195 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.230174 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lkgcd"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.246842 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvd6j\" (UniqueName: \"kubernetes.io/projected/b85ff85e-86bc-45c0-940e-09c62beba8e2-kube-api-access-pvd6j\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.248827 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.248922 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9msj\" (UniqueName: \"kubernetes.io/projected/9e6712c0-048c-4432-aad1-2dc9e280ae0b-kube-api-access-p9msj\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249012 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249123 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwhnx\" (UniqueName: \"kubernetes.io/projected/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-kube-api-access-dwhnx\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249228 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-logs\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249342 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249420 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc7b7d9-372e-4aa4-9776-f67e78497501-logs\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249489 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249688 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-config-data\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249790 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-config-data\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249863 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.249963 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.250047 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-config\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.250125 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-config-data\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.250201 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnm8w\" (UniqueName: \"kubernetes.io/projected/3dc7b7d9-372e-4aa4-9776-f67e78497501-kube-api-access-rnm8w\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.250272 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.250606 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc7b7d9-372e-4aa4-9776-f67e78497501-logs\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.254531 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-logs\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.264624 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-config-data\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.265789 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.266008 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.266372 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-config-data\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.267072 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-config-data\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.274567 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnm8w\" (UniqueName: \"kubernetes.io/projected/3dc7b7d9-372e-4aa4-9776-f67e78497501-kube-api-access-rnm8w\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.276501 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.290602 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9msj\" (UniqueName: \"kubernetes.io/projected/9e6712c0-048c-4432-aad1-2dc9e280ae0b-kube-api-access-p9msj\") pod \"nova-scheduler-0\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.298494 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.318331 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwhnx\" (UniqueName: \"kubernetes.io/projected/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-kube-api-access-dwhnx\") pod \"nova-api-0\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.351337 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.351392 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.351446 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-config\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.351482 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.351570 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvd6j\" (UniqueName: \"kubernetes.io/projected/b85ff85e-86bc-45c0-940e-09c62beba8e2-kube-api-access-pvd6j\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.351598 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.352640 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.353226 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.354052 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.354188 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.354299 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-config\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.376422 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvd6j\" (UniqueName: \"kubernetes.io/projected/b85ff85e-86bc-45c0-940e-09c62beba8e2-kube-api-access-pvd6j\") pod \"dnsmasq-dns-9b86998b5-lkgcd\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.410209 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.419416 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.431599 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.608496 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.872894 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-f52xr"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.895331 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.937441 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.955392 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f52xr" event={"ID":"c93d5fea-7f33-47a2-916a-80e9c68a89f2","Type":"ContainerStarted","Data":"9750d77e259a4bd1e3d97a825446a8774d3fb64408e98874a049baecfb35a4cf"} Nov 28 14:49:10 crc kubenswrapper[4817]: I1128 14:49:10.981850 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af","Type":"ContainerStarted","Data":"9b4dfb4ba412a8cdfcf21623f9f57e1d26816429ba57922d6ac74f91c78dbe34"} Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.113587 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.287796 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lkgcd"] Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.294255 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.503811 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.647811 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q994c"] Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.655453 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.657398 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.657748 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.671579 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q994c"] Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.706300 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.706375 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-config-data\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.706521 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-scripts\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.706555 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnslv\" (UniqueName: \"kubernetes.io/projected/1b2f8c40-8b05-4037-8add-58252a9f996d-kube-api-access-pnslv\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.807752 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-scripts\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.808024 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnslv\" (UniqueName: \"kubernetes.io/projected/1b2f8c40-8b05-4037-8add-58252a9f996d-kube-api-access-pnslv\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.808054 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.808118 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-config-data\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.812905 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-scripts\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.818765 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.823305 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-config-data\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.829215 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnslv\" (UniqueName: \"kubernetes.io/projected/1b2f8c40-8b05-4037-8add-58252a9f996d-kube-api-access-pnslv\") pod \"nova-cell1-conductor-db-sync-q994c\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:11 crc kubenswrapper[4817]: I1128 14:49:11.996023 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f52xr" event={"ID":"c93d5fea-7f33-47a2-916a-80e9c68a89f2","Type":"ContainerStarted","Data":"18e8370affb31e5bb70b35b5de7775b06051710af0ef2e5cc37bf860c16698f6"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:11.999154 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerStarted","Data":"c2d3804c26ea1e09e954d8877f509b0ce97f55b79c2ac5710c531c1f5d495fe3"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:11.999218 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.003699 4817 generic.go:334] "Generic (PLEG): container finished" podID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerID="c0e3885d125a5380218f976fa5fb5c0d354a9c86f89898a2160cb8da22bbe3a5" exitCode=0 Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.003741 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" event={"ID":"b85ff85e-86bc-45c0-940e-09c62beba8e2","Type":"ContainerDied","Data":"c0e3885d125a5380218f976fa5fb5c0d354a9c86f89898a2160cb8da22bbe3a5"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.003767 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" event={"ID":"b85ff85e-86bc-45c0-940e-09c62beba8e2","Type":"ContainerStarted","Data":"089ca6f76699531ed55c051ca44899f5cb3021e6017fced07c710549281bff53"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.006205 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3dc7b7d9-372e-4aa4-9776-f67e78497501","Type":"ContainerStarted","Data":"7725e1f08d4b1cff341d369b09ad8a84be3fc4c6d44694da3b7499d4efdbe696"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.010736 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e6712c0-048c-4432-aad1-2dc9e280ae0b","Type":"ContainerStarted","Data":"c638d5280c3cea9e88fa19ddcf1aa542eb14ab06a1bcde686b5896f4aa049224"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.012783 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b","Type":"ContainerStarted","Data":"099a7293d006ca2990d6c029fedf32236bb7afaefbd35e76461cf0b3118a426b"} Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.025875 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-f52xr" podStartSLOduration=3.025854455 podStartE2EDuration="3.025854455s" podCreationTimestamp="2025-11-28 14:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:12.018139394 +0000 UTC m=+1174.606117660" watchObservedRunningTime="2025-11-28 14:49:12.025854455 +0000 UTC m=+1174.613832721" Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.052561 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.063515 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.708721915 podStartE2EDuration="8.063497071s" podCreationTimestamp="2025-11-28 14:49:04 +0000 UTC" firstStartedPulling="2025-11-28 14:49:05.67940464 +0000 UTC m=+1168.267382916" lastFinishedPulling="2025-11-28 14:49:11.034179806 +0000 UTC m=+1173.622158072" observedRunningTime="2025-11-28 14:49:12.053259196 +0000 UTC m=+1174.641237462" watchObservedRunningTime="2025-11-28 14:49:12.063497071 +0000 UTC m=+1174.651475337" Nov 28 14:49:12 crc kubenswrapper[4817]: I1128 14:49:12.614904 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q994c"] Nov 28 14:49:13 crc kubenswrapper[4817]: I1128 14:49:13.035480 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" event={"ID":"b85ff85e-86bc-45c0-940e-09c62beba8e2","Type":"ContainerStarted","Data":"dc89bf855fd55526e5c99aa2850da18a9ca6372eb5bf070a2897f42a29cc90d1"} Nov 28 14:49:13 crc kubenswrapper[4817]: I1128 14:49:13.035644 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:13 crc kubenswrapper[4817]: I1128 14:49:13.038582 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q994c" event={"ID":"1b2f8c40-8b05-4037-8add-58252a9f996d","Type":"ContainerStarted","Data":"c0ba863b2241c0e64d5f8e0f2b6215fbcaafba2c8ab5101af74a4d7b5e5b3c60"} Nov 28 14:49:13 crc kubenswrapper[4817]: I1128 14:49:13.057239 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" podStartSLOduration=3.057080748 podStartE2EDuration="3.057080748s" podCreationTimestamp="2025-11-28 14:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:13.055901448 +0000 UTC m=+1175.643879734" watchObservedRunningTime="2025-11-28 14:49:13.057080748 +0000 UTC m=+1175.645059014" Nov 28 14:49:13 crc kubenswrapper[4817]: I1128 14:49:13.805713 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:13 crc kubenswrapper[4817]: I1128 14:49:13.823101 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.081379 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e6712c0-048c-4432-aad1-2dc9e280ae0b","Type":"ContainerStarted","Data":"ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.088092 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af","Type":"ContainerStarted","Data":"1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.088241 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66" gracePeriod=30 Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.105365 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b","Type":"ContainerStarted","Data":"8fceb98fa02b7b618460275ac1431733322159891f150c25896b4e6525121c9e"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.105691 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b","Type":"ContainerStarted","Data":"2a58f80eb43af69ff49d4f4c718fb00622027d52e25ee11b7ee7a6b36a80bf22"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.107601 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q994c" event={"ID":"1b2f8c40-8b05-4037-8add-58252a9f996d","Type":"ContainerStarted","Data":"53266733de55237f9fe7dbd93ec510da9273166899b424c157266d18d42e5df4"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.113243 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3dc7b7d9-372e-4aa4-9776-f67e78497501","Type":"ContainerStarted","Data":"922fe1307620f1d184e317194dfeb14ba342a6fcce42132b77ea64e698e2ce7c"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.113291 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3dc7b7d9-372e-4aa4-9776-f67e78497501","Type":"ContainerStarted","Data":"8895c07e57a64a48f203c27c3d85f872efe2f2c0e5ea2eb60cadd59a9215dae7"} Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.113413 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-log" containerID="cri-o://8895c07e57a64a48f203c27c3d85f872efe2f2c0e5ea2eb60cadd59a9215dae7" gracePeriod=30 Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.113712 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-metadata" containerID="cri-o://922fe1307620f1d184e317194dfeb14ba342a6fcce42132b77ea64e698e2ce7c" gracePeriod=30 Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.115826 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.810862379 podStartE2EDuration="6.115806845s" podCreationTimestamp="2025-11-28 14:49:10 +0000 UTC" firstStartedPulling="2025-11-28 14:49:11.429471497 +0000 UTC m=+1174.017449763" lastFinishedPulling="2025-11-28 14:49:14.734415963 +0000 UTC m=+1177.322394229" observedRunningTime="2025-11-28 14:49:16.109055288 +0000 UTC m=+1178.697033554" watchObservedRunningTime="2025-11-28 14:49:16.115806845 +0000 UTC m=+1178.703785121" Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.143769 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.888855818 podStartE2EDuration="7.14374646s" podCreationTimestamp="2025-11-28 14:49:09 +0000 UTC" firstStartedPulling="2025-11-28 14:49:11.4875376 +0000 UTC m=+1174.075515866" lastFinishedPulling="2025-11-28 14:49:14.742428242 +0000 UTC m=+1177.330406508" observedRunningTime="2025-11-28 14:49:16.133276339 +0000 UTC m=+1178.721254605" watchObservedRunningTime="2025-11-28 14:49:16.14374646 +0000 UTC m=+1178.731724726" Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.154786 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-q994c" podStartSLOduration=5.154769014 podStartE2EDuration="5.154769014s" podCreationTimestamp="2025-11-28 14:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:16.153052451 +0000 UTC m=+1178.741030717" watchObservedRunningTime="2025-11-28 14:49:16.154769014 +0000 UTC m=+1178.742747290" Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.169984 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.330738651 podStartE2EDuration="7.169970341s" podCreationTimestamp="2025-11-28 14:49:09 +0000 UTC" firstStartedPulling="2025-11-28 14:49:10.895134641 +0000 UTC m=+1173.483112907" lastFinishedPulling="2025-11-28 14:49:14.734366331 +0000 UTC m=+1177.322344597" observedRunningTime="2025-11-28 14:49:16.166830173 +0000 UTC m=+1178.754808439" watchObservedRunningTime="2025-11-28 14:49:16.169970341 +0000 UTC m=+1178.757948607" Nov 28 14:49:16 crc kubenswrapper[4817]: I1128 14:49:16.188698 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.68128168 podStartE2EDuration="7.188681556s" podCreationTimestamp="2025-11-28 14:49:09 +0000 UTC" firstStartedPulling="2025-11-28 14:49:11.226954975 +0000 UTC m=+1173.814933261" lastFinishedPulling="2025-11-28 14:49:14.734354871 +0000 UTC m=+1177.322333137" observedRunningTime="2025-11-28 14:49:16.183048546 +0000 UTC m=+1178.771026812" watchObservedRunningTime="2025-11-28 14:49:16.188681556 +0000 UTC m=+1178.776659822" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.132072 4817 generic.go:334] "Generic (PLEG): container finished" podID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerID="922fe1307620f1d184e317194dfeb14ba342a6fcce42132b77ea64e698e2ce7c" exitCode=0 Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.132740 4817 generic.go:334] "Generic (PLEG): container finished" podID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerID="8895c07e57a64a48f203c27c3d85f872efe2f2c0e5ea2eb60cadd59a9215dae7" exitCode=143 Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.132152 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3dc7b7d9-372e-4aa4-9776-f67e78497501","Type":"ContainerDied","Data":"922fe1307620f1d184e317194dfeb14ba342a6fcce42132b77ea64e698e2ce7c"} Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.132802 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3dc7b7d9-372e-4aa4-9776-f67e78497501","Type":"ContainerDied","Data":"8895c07e57a64a48f203c27c3d85f872efe2f2c0e5ea2eb60cadd59a9215dae7"} Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.284815 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.332345 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-config-data\") pod \"3dc7b7d9-372e-4aa4-9776-f67e78497501\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.332389 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc7b7d9-372e-4aa4-9776-f67e78497501-logs\") pod \"3dc7b7d9-372e-4aa4-9776-f67e78497501\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.332469 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-combined-ca-bundle\") pod \"3dc7b7d9-372e-4aa4-9776-f67e78497501\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.332606 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnm8w\" (UniqueName: \"kubernetes.io/projected/3dc7b7d9-372e-4aa4-9776-f67e78497501-kube-api-access-rnm8w\") pod \"3dc7b7d9-372e-4aa4-9776-f67e78497501\" (UID: \"3dc7b7d9-372e-4aa4-9776-f67e78497501\") " Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.334397 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc7b7d9-372e-4aa4-9776-f67e78497501-logs" (OuterVolumeSpecName: "logs") pod "3dc7b7d9-372e-4aa4-9776-f67e78497501" (UID: "3dc7b7d9-372e-4aa4-9776-f67e78497501"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.358539 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc7b7d9-372e-4aa4-9776-f67e78497501-kube-api-access-rnm8w" (OuterVolumeSpecName: "kube-api-access-rnm8w") pod "3dc7b7d9-372e-4aa4-9776-f67e78497501" (UID: "3dc7b7d9-372e-4aa4-9776-f67e78497501"). InnerVolumeSpecName "kube-api-access-rnm8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.368808 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-config-data" (OuterVolumeSpecName: "config-data") pod "3dc7b7d9-372e-4aa4-9776-f67e78497501" (UID: "3dc7b7d9-372e-4aa4-9776-f67e78497501"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.394794 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dc7b7d9-372e-4aa4-9776-f67e78497501" (UID: "3dc7b7d9-372e-4aa4-9776-f67e78497501"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.434979 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.435014 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc7b7d9-372e-4aa4-9776-f67e78497501-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.435024 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc7b7d9-372e-4aa4-9776-f67e78497501-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:17 crc kubenswrapper[4817]: I1128 14:49:17.435036 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnm8w\" (UniqueName: \"kubernetes.io/projected/3dc7b7d9-372e-4aa4-9776-f67e78497501-kube-api-access-rnm8w\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.142129 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3dc7b7d9-372e-4aa4-9776-f67e78497501","Type":"ContainerDied","Data":"7725e1f08d4b1cff341d369b09ad8a84be3fc4c6d44694da3b7499d4efdbe696"} Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.142392 4817 scope.go:117] "RemoveContainer" containerID="922fe1307620f1d184e317194dfeb14ba342a6fcce42132b77ea64e698e2ce7c" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.143089 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.169994 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.177217 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.182805 4817 scope.go:117] "RemoveContainer" containerID="8895c07e57a64a48f203c27c3d85f872efe2f2c0e5ea2eb60cadd59a9215dae7" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.214324 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:18 crc kubenswrapper[4817]: E1128 14:49:18.214700 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-metadata" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.214713 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-metadata" Nov 28 14:49:18 crc kubenswrapper[4817]: E1128 14:49:18.214744 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-log" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.214750 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-log" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.214932 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-metadata" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.214957 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" containerName="nova-metadata-log" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.215878 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.219090 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.223181 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.243309 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.254935 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.254976 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b70eb86e-af36-460b-a474-61bf82e86dd9-logs\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.255097 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.255115 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-config-data\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.255158 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nzzw\" (UniqueName: \"kubernetes.io/projected/b70eb86e-af36-460b-a474-61bf82e86dd9-kube-api-access-8nzzw\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.357703 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nzzw\" (UniqueName: \"kubernetes.io/projected/b70eb86e-af36-460b-a474-61bf82e86dd9-kube-api-access-8nzzw\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.357784 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.357820 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b70eb86e-af36-460b-a474-61bf82e86dd9-logs\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.357969 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.357992 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-config-data\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.358662 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b70eb86e-af36-460b-a474-61bf82e86dd9-logs\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.376645 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-config-data\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.388495 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.389563 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.393210 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nzzw\" (UniqueName: \"kubernetes.io/projected/b70eb86e-af36-460b-a474-61bf82e86dd9-kube-api-access-8nzzw\") pod \"nova-metadata-0\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " pod="openstack/nova-metadata-0" Nov 28 14:49:18 crc kubenswrapper[4817]: I1128 14:49:18.535261 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:19 crc kubenswrapper[4817]: I1128 14:49:19.085567 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:19 crc kubenswrapper[4817]: I1128 14:49:19.152585 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b70eb86e-af36-460b-a474-61bf82e86dd9","Type":"ContainerStarted","Data":"595eb84df10408524f07711b540781196deee9dad9f6ec654a2bc4dd90db95fa"} Nov 28 14:49:19 crc kubenswrapper[4817]: I1128 14:49:19.765312 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dc7b7d9-372e-4aa4-9776-f67e78497501" path="/var/lib/kubelet/pods/3dc7b7d9-372e-4aa4-9776-f67e78497501/volumes" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.166490 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b70eb86e-af36-460b-a474-61bf82e86dd9","Type":"ContainerStarted","Data":"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e"} Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.166538 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b70eb86e-af36-460b-a474-61bf82e86dd9","Type":"ContainerStarted","Data":"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e"} Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.168897 4817 generic.go:334] "Generic (PLEG): container finished" podID="c93d5fea-7f33-47a2-916a-80e9c68a89f2" containerID="18e8370affb31e5bb70b35b5de7775b06051710af0ef2e5cc37bf860c16698f6" exitCode=0 Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.168950 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f52xr" event={"ID":"c93d5fea-7f33-47a2-916a-80e9c68a89f2","Type":"ContainerDied","Data":"18e8370affb31e5bb70b35b5de7775b06051710af0ef2e5cc37bf860c16698f6"} Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.190892 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.190859175 podStartE2EDuration="2.190859175s" podCreationTimestamp="2025-11-28 14:49:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:20.183636586 +0000 UTC m=+1182.771614882" watchObservedRunningTime="2025-11-28 14:49:20.190859175 +0000 UTC m=+1182.778837491" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.299956 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.420225 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.420274 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.434189 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.465557 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.543990 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-cxp27"] Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.544680 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerName="dnsmasq-dns" containerID="cri-o://35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739" gracePeriod=10 Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.609385 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 14:49:20 crc kubenswrapper[4817]: I1128 14:49:20.609885 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.066579 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.124224 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-nb\") pod \"610bb394-f0ff-4876-bf71-86aacfa8f690\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.124282 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-sb\") pod \"610bb394-f0ff-4876-bf71-86aacfa8f690\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.124343 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-svc\") pod \"610bb394-f0ff-4876-bf71-86aacfa8f690\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.125042 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-config\") pod \"610bb394-f0ff-4876-bf71-86aacfa8f690\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.125103 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-swift-storage-0\") pod \"610bb394-f0ff-4876-bf71-86aacfa8f690\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.125176 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcfk5\" (UniqueName: \"kubernetes.io/projected/610bb394-f0ff-4876-bf71-86aacfa8f690-kube-api-access-wcfk5\") pod \"610bb394-f0ff-4876-bf71-86aacfa8f690\" (UID: \"610bb394-f0ff-4876-bf71-86aacfa8f690\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.151354 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/610bb394-f0ff-4876-bf71-86aacfa8f690-kube-api-access-wcfk5" (OuterVolumeSpecName: "kube-api-access-wcfk5") pod "610bb394-f0ff-4876-bf71-86aacfa8f690" (UID: "610bb394-f0ff-4876-bf71-86aacfa8f690"). InnerVolumeSpecName "kube-api-access-wcfk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.182112 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "610bb394-f0ff-4876-bf71-86aacfa8f690" (UID: "610bb394-f0ff-4876-bf71-86aacfa8f690"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.193345 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "610bb394-f0ff-4876-bf71-86aacfa8f690" (UID: "610bb394-f0ff-4876-bf71-86aacfa8f690"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.194562 4817 generic.go:334] "Generic (PLEG): container finished" podID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerID="35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739" exitCode=0 Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.194654 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.195024 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" event={"ID":"610bb394-f0ff-4876-bf71-86aacfa8f690","Type":"ContainerDied","Data":"35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739"} Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.195056 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-cxp27" event={"ID":"610bb394-f0ff-4876-bf71-86aacfa8f690","Type":"ContainerDied","Data":"037990dc8156ba9da45f18cb7e61b330a964a5f2c9d815684ce22b0a9b209295"} Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.195072 4817 scope.go:117] "RemoveContainer" containerID="35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.196028 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-config" (OuterVolumeSpecName: "config") pod "610bb394-f0ff-4876-bf71-86aacfa8f690" (UID: "610bb394-f0ff-4876-bf71-86aacfa8f690"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.212216 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "610bb394-f0ff-4876-bf71-86aacfa8f690" (UID: "610bb394-f0ff-4876-bf71-86aacfa8f690"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.234308 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.240899 4817 scope.go:117] "RemoveContainer" containerID="a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.241416 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "610bb394-f0ff-4876-bf71-86aacfa8f690" (UID: "610bb394-f0ff-4876-bf71-86aacfa8f690"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.242578 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.242694 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.242796 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcfk5\" (UniqueName: \"kubernetes.io/projected/610bb394-f0ff-4876-bf71-86aacfa8f690-kube-api-access-wcfk5\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.242864 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.242941 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.243020 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/610bb394-f0ff-4876-bf71-86aacfa8f690-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.274951 4817 scope.go:117] "RemoveContainer" containerID="35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739" Nov 28 14:49:21 crc kubenswrapper[4817]: E1128 14:49:21.277851 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739\": container with ID starting with 35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739 not found: ID does not exist" containerID="35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.278065 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739"} err="failed to get container status \"35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739\": rpc error: code = NotFound desc = could not find container \"35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739\": container with ID starting with 35cac2979ed074a7711a148349c56ef3a9d32c6b366b8a3cee4d06634e27c739 not found: ID does not exist" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.278164 4817 scope.go:117] "RemoveContainer" containerID="a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7" Nov 28 14:49:21 crc kubenswrapper[4817]: E1128 14:49:21.278934 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7\": container with ID starting with a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7 not found: ID does not exist" containerID="a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.278974 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7"} err="failed to get container status \"a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7\": rpc error: code = NotFound desc = could not find container \"a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7\": container with ID starting with a5191c74cc3fd23659e8df91244c5dd878184b25ef19cc6355f30fbf557238b7 not found: ID does not exist" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.534474 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-cxp27"] Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.534978 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.545810 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-cxp27"] Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.650814 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-combined-ca-bundle\") pod \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.651151 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-config-data\") pod \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.651275 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq2m5\" (UniqueName: \"kubernetes.io/projected/c93d5fea-7f33-47a2-916a-80e9c68a89f2-kube-api-access-tq2m5\") pod \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.651376 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-scripts\") pod \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\" (UID: \"c93d5fea-7f33-47a2-916a-80e9c68a89f2\") " Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.652230 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.652454 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.660076 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c93d5fea-7f33-47a2-916a-80e9c68a89f2-kube-api-access-tq2m5" (OuterVolumeSpecName: "kube-api-access-tq2m5") pod "c93d5fea-7f33-47a2-916a-80e9c68a89f2" (UID: "c93d5fea-7f33-47a2-916a-80e9c68a89f2"). InnerVolumeSpecName "kube-api-access-tq2m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.678462 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-config-data" (OuterVolumeSpecName: "config-data") pod "c93d5fea-7f33-47a2-916a-80e9c68a89f2" (UID: "c93d5fea-7f33-47a2-916a-80e9c68a89f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.684617 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c93d5fea-7f33-47a2-916a-80e9c68a89f2" (UID: "c93d5fea-7f33-47a2-916a-80e9c68a89f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.696582 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-scripts" (OuterVolumeSpecName: "scripts") pod "c93d5fea-7f33-47a2-916a-80e9c68a89f2" (UID: "c93d5fea-7f33-47a2-916a-80e9c68a89f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.749330 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" path="/var/lib/kubelet/pods/610bb394-f0ff-4876-bf71-86aacfa8f690/volumes" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.753793 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq2m5\" (UniqueName: \"kubernetes.io/projected/c93d5fea-7f33-47a2-916a-80e9c68a89f2-kube-api-access-tq2m5\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.753836 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.753847 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:21 crc kubenswrapper[4817]: I1128 14:49:21.753857 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93d5fea-7f33-47a2-916a-80e9c68a89f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.227449 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-f52xr" Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.228041 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-f52xr" event={"ID":"c93d5fea-7f33-47a2-916a-80e9c68a89f2","Type":"ContainerDied","Data":"9750d77e259a4bd1e3d97a825446a8774d3fb64408e98874a049baecfb35a4cf"} Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.228065 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9750d77e259a4bd1e3d97a825446a8774d3fb64408e98874a049baecfb35a4cf" Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.451448 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.451696 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-log" containerID="cri-o://2a58f80eb43af69ff49d4f4c718fb00622027d52e25ee11b7ee7a6b36a80bf22" gracePeriod=30 Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.451762 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-api" containerID="cri-o://8fceb98fa02b7b618460275ac1431733322159891f150c25896b4e6525121c9e" gracePeriod=30 Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.467574 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.499058 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.499291 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-log" containerID="cri-o://5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e" gracePeriod=30 Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.499618 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-metadata" containerID="cri-o://8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e" gracePeriod=30 Nov 28 14:49:22 crc kubenswrapper[4817]: I1128 14:49:22.987294 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.078784 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-config-data\") pod \"b70eb86e-af36-460b-a474-61bf82e86dd9\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.078894 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-nova-metadata-tls-certs\") pod \"b70eb86e-af36-460b-a474-61bf82e86dd9\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.078921 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-combined-ca-bundle\") pod \"b70eb86e-af36-460b-a474-61bf82e86dd9\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.079022 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b70eb86e-af36-460b-a474-61bf82e86dd9-logs\") pod \"b70eb86e-af36-460b-a474-61bf82e86dd9\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.079096 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nzzw\" (UniqueName: \"kubernetes.io/projected/b70eb86e-af36-460b-a474-61bf82e86dd9-kube-api-access-8nzzw\") pod \"b70eb86e-af36-460b-a474-61bf82e86dd9\" (UID: \"b70eb86e-af36-460b-a474-61bf82e86dd9\") " Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.079766 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70eb86e-af36-460b-a474-61bf82e86dd9-logs" (OuterVolumeSpecName: "logs") pod "b70eb86e-af36-460b-a474-61bf82e86dd9" (UID: "b70eb86e-af36-460b-a474-61bf82e86dd9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.084666 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70eb86e-af36-460b-a474-61bf82e86dd9-kube-api-access-8nzzw" (OuterVolumeSpecName: "kube-api-access-8nzzw") pod "b70eb86e-af36-460b-a474-61bf82e86dd9" (UID: "b70eb86e-af36-460b-a474-61bf82e86dd9"). InnerVolumeSpecName "kube-api-access-8nzzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.110305 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b70eb86e-af36-460b-a474-61bf82e86dd9" (UID: "b70eb86e-af36-460b-a474-61bf82e86dd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.121354 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-config-data" (OuterVolumeSpecName: "config-data") pod "b70eb86e-af36-460b-a474-61bf82e86dd9" (UID: "b70eb86e-af36-460b-a474-61bf82e86dd9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.141813 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b70eb86e-af36-460b-a474-61bf82e86dd9" (UID: "b70eb86e-af36-460b-a474-61bf82e86dd9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.181392 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.181432 4817 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.181448 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70eb86e-af36-460b-a474-61bf82e86dd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.181461 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b70eb86e-af36-460b-a474-61bf82e86dd9-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.181472 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nzzw\" (UniqueName: \"kubernetes.io/projected/b70eb86e-af36-460b-a474-61bf82e86dd9-kube-api-access-8nzzw\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237240 4817 generic.go:334] "Generic (PLEG): container finished" podID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerID="8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e" exitCode=0 Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237267 4817 generic.go:334] "Generic (PLEG): container finished" podID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerID="5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e" exitCode=143 Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237303 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b70eb86e-af36-460b-a474-61bf82e86dd9","Type":"ContainerDied","Data":"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e"} Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237329 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b70eb86e-af36-460b-a474-61bf82e86dd9","Type":"ContainerDied","Data":"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e"} Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237339 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b70eb86e-af36-460b-a474-61bf82e86dd9","Type":"ContainerDied","Data":"595eb84df10408524f07711b540781196deee9dad9f6ec654a2bc4dd90db95fa"} Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237354 4817 scope.go:117] "RemoveContainer" containerID="8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.237463 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.241347 4817 generic.go:334] "Generic (PLEG): container finished" podID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerID="2a58f80eb43af69ff49d4f4c718fb00622027d52e25ee11b7ee7a6b36a80bf22" exitCode=143 Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.241462 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b","Type":"ContainerDied","Data":"2a58f80eb43af69ff49d4f4c718fb00622027d52e25ee11b7ee7a6b36a80bf22"} Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.241479 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" containerName="nova-scheduler-scheduler" containerID="cri-o://ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" gracePeriod=30 Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.259404 4817 scope.go:117] "RemoveContainer" containerID="5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.275615 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.297406 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.299526 4817 scope.go:117] "RemoveContainer" containerID="8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e" Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.300068 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e\": container with ID starting with 8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e not found: ID does not exist" containerID="8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.300108 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e"} err="failed to get container status \"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e\": rpc error: code = NotFound desc = could not find container \"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e\": container with ID starting with 8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e not found: ID does not exist" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.300136 4817 scope.go:117] "RemoveContainer" containerID="5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e" Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.300553 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e\": container with ID starting with 5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e not found: ID does not exist" containerID="5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.300582 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e"} err="failed to get container status \"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e\": rpc error: code = NotFound desc = could not find container \"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e\": container with ID starting with 5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e not found: ID does not exist" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.300608 4817 scope.go:117] "RemoveContainer" containerID="8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.301145 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e"} err="failed to get container status \"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e\": rpc error: code = NotFound desc = could not find container \"8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e\": container with ID starting with 8cf6519e2b6328dbb0c99fa4dd7f3c68e0d2a094a72f211ef3a9d61490b7589e not found: ID does not exist" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.301167 4817 scope.go:117] "RemoveContainer" containerID="5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.301455 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e"} err="failed to get container status \"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e\": rpc error: code = NotFound desc = could not find container \"5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e\": container with ID starting with 5b05214f9d66dcf0c1709a7082d62f5f7990d0d69af527cb4622f40621f91c8e not found: ID does not exist" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.324751 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.325151 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-metadata" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325166 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-metadata" Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.325194 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93d5fea-7f33-47a2-916a-80e9c68a89f2" containerName="nova-manage" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325201 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93d5fea-7f33-47a2-916a-80e9c68a89f2" containerName="nova-manage" Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.325217 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-log" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325223 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-log" Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.325237 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerName="dnsmasq-dns" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325243 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerName="dnsmasq-dns" Nov 28 14:49:23 crc kubenswrapper[4817]: E1128 14:49:23.325256 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerName="init" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325264 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerName="init" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325424 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="610bb394-f0ff-4876-bf71-86aacfa8f690" containerName="dnsmasq-dns" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325437 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-log" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325453 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c93d5fea-7f33-47a2-916a-80e9c68a89f2" containerName="nova-manage" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.325465 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" containerName="nova-metadata-metadata" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.326647 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.335643 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.335665 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.381357 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.489328 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6290a311-8ee7-4864-b63f-a6a5225c0df4-logs\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.489397 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgxgz\" (UniqueName: \"kubernetes.io/projected/6290a311-8ee7-4864-b63f-a6a5225c0df4-kube-api-access-fgxgz\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.489435 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.489680 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-config-data\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.489794 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.591252 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6290a311-8ee7-4864-b63f-a6a5225c0df4-logs\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.591335 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgxgz\" (UniqueName: \"kubernetes.io/projected/6290a311-8ee7-4864-b63f-a6a5225c0df4-kube-api-access-fgxgz\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.591386 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.591489 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-config-data\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.591568 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.591671 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6290a311-8ee7-4864-b63f-a6a5225c0df4-logs\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.596217 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.597872 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.600182 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-config-data\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.610112 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgxgz\" (UniqueName: \"kubernetes.io/projected/6290a311-8ee7-4864-b63f-a6a5225c0df4-kube-api-access-fgxgz\") pod \"nova-metadata-0\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.653131 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:49:23 crc kubenswrapper[4817]: I1128 14:49:23.759892 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b70eb86e-af36-460b-a474-61bf82e86dd9" path="/var/lib/kubelet/pods/b70eb86e-af36-460b-a474-61bf82e86dd9/volumes" Nov 28 14:49:24 crc kubenswrapper[4817]: I1128 14:49:24.045184 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:49:24 crc kubenswrapper[4817]: I1128 14:49:24.045273 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:49:24 crc kubenswrapper[4817]: I1128 14:49:24.140684 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:49:24 crc kubenswrapper[4817]: I1128 14:49:24.253322 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6290a311-8ee7-4864-b63f-a6a5225c0df4","Type":"ContainerStarted","Data":"1c6067aeac80518fd0e293d695a751370636c3c149566fd2a33a13441c9d76cd"} Nov 28 14:49:24 crc kubenswrapper[4817]: I1128 14:49:24.256313 4817 generic.go:334] "Generic (PLEG): container finished" podID="1b2f8c40-8b05-4037-8add-58252a9f996d" containerID="53266733de55237f9fe7dbd93ec510da9273166899b424c157266d18d42e5df4" exitCode=0 Nov 28 14:49:24 crc kubenswrapper[4817]: I1128 14:49:24.256356 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q994c" event={"ID":"1b2f8c40-8b05-4037-8add-58252a9f996d","Type":"ContainerDied","Data":"53266733de55237f9fe7dbd93ec510da9273166899b424c157266d18d42e5df4"} Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.284076 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6290a311-8ee7-4864-b63f-a6a5225c0df4","Type":"ContainerStarted","Data":"5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0"} Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.284383 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6290a311-8ee7-4864-b63f-a6a5225c0df4","Type":"ContainerStarted","Data":"15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d"} Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.321394 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.321369029 podStartE2EDuration="2.321369029s" podCreationTimestamp="2025-11-28 14:49:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:25.301525555 +0000 UTC m=+1187.889503851" watchObservedRunningTime="2025-11-28 14:49:25.321369029 +0000 UTC m=+1187.909347305" Nov 28 14:49:25 crc kubenswrapper[4817]: E1128 14:49:25.421921 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 14:49:25 crc kubenswrapper[4817]: E1128 14:49:25.425175 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 14:49:25 crc kubenswrapper[4817]: E1128 14:49:25.426185 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 14:49:25 crc kubenswrapper[4817]: E1128 14:49:25.426221 4817 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" containerName="nova-scheduler-scheduler" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.646214 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.760405 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnslv\" (UniqueName: \"kubernetes.io/projected/1b2f8c40-8b05-4037-8add-58252a9f996d-kube-api-access-pnslv\") pod \"1b2f8c40-8b05-4037-8add-58252a9f996d\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.760937 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-config-data\") pod \"1b2f8c40-8b05-4037-8add-58252a9f996d\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.761372 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-scripts\") pod \"1b2f8c40-8b05-4037-8add-58252a9f996d\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.761600 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-combined-ca-bundle\") pod \"1b2f8c40-8b05-4037-8add-58252a9f996d\" (UID: \"1b2f8c40-8b05-4037-8add-58252a9f996d\") " Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.766267 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-scripts" (OuterVolumeSpecName: "scripts") pod "1b2f8c40-8b05-4037-8add-58252a9f996d" (UID: "1b2f8c40-8b05-4037-8add-58252a9f996d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.766326 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2f8c40-8b05-4037-8add-58252a9f996d-kube-api-access-pnslv" (OuterVolumeSpecName: "kube-api-access-pnslv") pod "1b2f8c40-8b05-4037-8add-58252a9f996d" (UID: "1b2f8c40-8b05-4037-8add-58252a9f996d"). InnerVolumeSpecName "kube-api-access-pnslv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.792792 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-config-data" (OuterVolumeSpecName: "config-data") pod "1b2f8c40-8b05-4037-8add-58252a9f996d" (UID: "1b2f8c40-8b05-4037-8add-58252a9f996d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.795587 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b2f8c40-8b05-4037-8add-58252a9f996d" (UID: "1b2f8c40-8b05-4037-8add-58252a9f996d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.864768 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.864804 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.864815 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f8c40-8b05-4037-8add-58252a9f996d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:25 crc kubenswrapper[4817]: I1128 14:49:25.864826 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnslv\" (UniqueName: \"kubernetes.io/projected/1b2f8c40-8b05-4037-8add-58252a9f996d-kube-api-access-pnslv\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.294254 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q994c" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.294246 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q994c" event={"ID":"1b2f8c40-8b05-4037-8add-58252a9f996d","Type":"ContainerDied","Data":"c0ba863b2241c0e64d5f8e0f2b6215fbcaafba2c8ab5101af74a4d7b5e5b3c60"} Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.295342 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0ba863b2241c0e64d5f8e0f2b6215fbcaafba2c8ab5101af74a4d7b5e5b3c60" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.357435 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 14:49:26 crc kubenswrapper[4817]: E1128 14:49:26.358099 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2f8c40-8b05-4037-8add-58252a9f996d" containerName="nova-cell1-conductor-db-sync" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.358116 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2f8c40-8b05-4037-8add-58252a9f996d" containerName="nova-cell1-conductor-db-sync" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.358332 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2f8c40-8b05-4037-8add-58252a9f996d" containerName="nova-cell1-conductor-db-sync" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.359016 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.361366 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.380595 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.476928 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37355c0f-d765-472c-b5d6-beb57bcf9869-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.477087 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np7x8\" (UniqueName: \"kubernetes.io/projected/37355c0f-d765-472c-b5d6-beb57bcf9869-kube-api-access-np7x8\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.477184 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37355c0f-d765-472c-b5d6-beb57bcf9869-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.579499 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37355c0f-d765-472c-b5d6-beb57bcf9869-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.579815 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np7x8\" (UniqueName: \"kubernetes.io/projected/37355c0f-d765-472c-b5d6-beb57bcf9869-kube-api-access-np7x8\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.579935 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37355c0f-d765-472c-b5d6-beb57bcf9869-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.598918 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37355c0f-d765-472c-b5d6-beb57bcf9869-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.598956 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37355c0f-d765-472c-b5d6-beb57bcf9869-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.608884 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np7x8\" (UniqueName: \"kubernetes.io/projected/37355c0f-d765-472c-b5d6-beb57bcf9869-kube-api-access-np7x8\") pod \"nova-cell1-conductor-0\" (UID: \"37355c0f-d765-472c-b5d6-beb57bcf9869\") " pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:26 crc kubenswrapper[4817]: I1128 14:49:26.683538 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.177362 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.179421 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.293129 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-config-data\") pod \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.293283 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-combined-ca-bundle\") pod \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.293318 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9msj\" (UniqueName: \"kubernetes.io/projected/9e6712c0-048c-4432-aad1-2dc9e280ae0b-kube-api-access-p9msj\") pod \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\" (UID: \"9e6712c0-048c-4432-aad1-2dc9e280ae0b\") " Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.296966 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6712c0-048c-4432-aad1-2dc9e280ae0b-kube-api-access-p9msj" (OuterVolumeSpecName: "kube-api-access-p9msj") pod "9e6712c0-048c-4432-aad1-2dc9e280ae0b" (UID: "9e6712c0-048c-4432-aad1-2dc9e280ae0b"). InnerVolumeSpecName "kube-api-access-p9msj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.303392 4817 generic.go:334] "Generic (PLEG): container finished" podID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" exitCode=0 Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.303437 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e6712c0-048c-4432-aad1-2dc9e280ae0b","Type":"ContainerDied","Data":"ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a"} Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.303456 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.303479 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e6712c0-048c-4432-aad1-2dc9e280ae0b","Type":"ContainerDied","Data":"c638d5280c3cea9e88fa19ddcf1aa542eb14ab06a1bcde686b5896f4aa049224"} Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.303526 4817 scope.go:117] "RemoveContainer" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.304525 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"37355c0f-d765-472c-b5d6-beb57bcf9869","Type":"ContainerStarted","Data":"491172abeb916300d8e3bf28c747157a48994342ba8474322f81856e66da8f4a"} Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.321998 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-config-data" (OuterVolumeSpecName: "config-data") pod "9e6712c0-048c-4432-aad1-2dc9e280ae0b" (UID: "9e6712c0-048c-4432-aad1-2dc9e280ae0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.322021 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e6712c0-048c-4432-aad1-2dc9e280ae0b" (UID: "9e6712c0-048c-4432-aad1-2dc9e280ae0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.326657 4817 scope.go:117] "RemoveContainer" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" Nov 28 14:49:27 crc kubenswrapper[4817]: E1128 14:49:27.327249 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a\": container with ID starting with ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a not found: ID does not exist" containerID="ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.327292 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a"} err="failed to get container status \"ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a\": rpc error: code = NotFound desc = could not find container \"ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a\": container with ID starting with ef589dc60ffd7e1db64d754480efe032b63efe9745c5ab96e17ffb754fd6c49a not found: ID does not exist" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.395911 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.396298 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6712c0-048c-4432-aad1-2dc9e280ae0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.396312 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9msj\" (UniqueName: \"kubernetes.io/projected/9e6712c0-048c-4432-aad1-2dc9e280ae0b-kube-api-access-p9msj\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.662137 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.674799 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.682885 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:27 crc kubenswrapper[4817]: E1128 14:49:27.683323 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" containerName="nova-scheduler-scheduler" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.683340 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" containerName="nova-scheduler-scheduler" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.683551 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" containerName="nova-scheduler-scheduler" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.684227 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.689484 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.708050 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.748647 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6712c0-048c-4432-aad1-2dc9e280ae0b" path="/var/lib/kubelet/pods/9e6712c0-048c-4432-aad1-2dc9e280ae0b/volumes" Nov 28 14:49:27 crc kubenswrapper[4817]: E1128 14:49:27.749558 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e6712c0_048c_4432_aad1_2dc9e280ae0b.slice\": RecentStats: unable to find data in memory cache]" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.809690 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.809765 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crb5g\" (UniqueName: \"kubernetes.io/projected/dd202581-e588-42a1-bd75-7ea6a8e03b23-kube-api-access-crb5g\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.809785 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-config-data\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.911352 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.911417 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crb5g\" (UniqueName: \"kubernetes.io/projected/dd202581-e588-42a1-bd75-7ea6a8e03b23-kube-api-access-crb5g\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.911442 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-config-data\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.915807 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-config-data\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.921614 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:27 crc kubenswrapper[4817]: I1128 14:49:27.929339 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crb5g\" (UniqueName: \"kubernetes.io/projected/dd202581-e588-42a1-bd75-7ea6a8e03b23-kube-api-access-crb5g\") pod \"nova-scheduler-0\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " pod="openstack/nova-scheduler-0" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.008710 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.319561 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"37355c0f-d765-472c-b5d6-beb57bcf9869","Type":"ContainerStarted","Data":"11b79bbdeb65f0b9010fd0fc674afb3146c9e4ec9065afbb2fddf090970ee8d8"} Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.320190 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.325044 4817 generic.go:334] "Generic (PLEG): container finished" podID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerID="8fceb98fa02b7b618460275ac1431733322159891f150c25896b4e6525121c9e" exitCode=0 Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.325096 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b","Type":"ContainerDied","Data":"8fceb98fa02b7b618460275ac1431733322159891f150c25896b4e6525121c9e"} Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.345127 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.345106107 podStartE2EDuration="2.345106107s" podCreationTimestamp="2025-11-28 14:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:28.336881242 +0000 UTC m=+1190.924859508" watchObservedRunningTime="2025-11-28 14:49:28.345106107 +0000 UTC m=+1190.933084373" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.426558 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.522624 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwhnx\" (UniqueName: \"kubernetes.io/projected/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-kube-api-access-dwhnx\") pod \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.522830 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-config-data\") pod \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.522910 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-combined-ca-bundle\") pod \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.522949 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-logs\") pod \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\" (UID: \"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b\") " Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.523751 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-logs" (OuterVolumeSpecName: "logs") pod "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" (UID: "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.533022 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-kube-api-access-dwhnx" (OuterVolumeSpecName: "kube-api-access-dwhnx") pod "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" (UID: "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b"). InnerVolumeSpecName "kube-api-access-dwhnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.595937 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-config-data" (OuterVolumeSpecName: "config-data") pod "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" (UID: "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.612059 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" (UID: "cc0d8d67-b5e2-4d76-8c5f-a5729406e57b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.633268 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.633291 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.633302 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.633310 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwhnx\" (UniqueName: \"kubernetes.io/projected/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b-kube-api-access-dwhnx\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.653728 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.654053 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 14:49:28 crc kubenswrapper[4817]: I1128 14:49:28.711607 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.334939 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.334943 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc0d8d67-b5e2-4d76-8c5f-a5729406e57b","Type":"ContainerDied","Data":"099a7293d006ca2990d6c029fedf32236bb7afaefbd35e76461cf0b3118a426b"} Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.335637 4817 scope.go:117] "RemoveContainer" containerID="8fceb98fa02b7b618460275ac1431733322159891f150c25896b4e6525121c9e" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.349622 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd202581-e588-42a1-bd75-7ea6a8e03b23","Type":"ContainerStarted","Data":"167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420"} Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.349655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd202581-e588-42a1-bd75-7ea6a8e03b23","Type":"ContainerStarted","Data":"4b4f742a0f4d18cf9599926b45a95c70ed32a55dae2aae46e1c8a5d6c30a812c"} Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.372465 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.372447093 podStartE2EDuration="2.372447093s" podCreationTimestamp="2025-11-28 14:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:29.370417422 +0000 UTC m=+1191.958395688" watchObservedRunningTime="2025-11-28 14:49:29.372447093 +0000 UTC m=+1191.960425359" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.376980 4817 scope.go:117] "RemoveContainer" containerID="2a58f80eb43af69ff49d4f4c718fb00622027d52e25ee11b7ee7a6b36a80bf22" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.397917 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.405664 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.415315 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:29 crc kubenswrapper[4817]: E1128 14:49:29.415879 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-api" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.415904 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-api" Nov 28 14:49:29 crc kubenswrapper[4817]: E1128 14:49:29.415923 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-log" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.415931 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-log" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.416938 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-log" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.416976 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" containerName="nova-api-api" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.418278 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.421938 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.432306 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.550860 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.550975 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-config-data\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.551044 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dc8bf0-06a6-4188-bee7-78a1c83374dd-logs\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.551164 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m42sg\" (UniqueName: \"kubernetes.io/projected/45dc8bf0-06a6-4188-bee7-78a1c83374dd-kube-api-access-m42sg\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.653254 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dc8bf0-06a6-4188-bee7-78a1c83374dd-logs\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.653399 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m42sg\" (UniqueName: \"kubernetes.io/projected/45dc8bf0-06a6-4188-bee7-78a1c83374dd-kube-api-access-m42sg\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.653436 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.653498 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-config-data\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.656557 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dc8bf0-06a6-4188-bee7-78a1c83374dd-logs\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.666510 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-config-data\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.670378 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.677755 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m42sg\" (UniqueName: \"kubernetes.io/projected/45dc8bf0-06a6-4188-bee7-78a1c83374dd-kube-api-access-m42sg\") pod \"nova-api-0\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.738266 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:29 crc kubenswrapper[4817]: I1128 14:49:29.747792 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0d8d67-b5e2-4d76-8c5f-a5729406e57b" path="/var/lib/kubelet/pods/cc0d8d67-b5e2-4d76-8c5f-a5729406e57b/volumes" Nov 28 14:49:30 crc kubenswrapper[4817]: I1128 14:49:30.233493 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:30 crc kubenswrapper[4817]: I1128 14:49:30.361191 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45dc8bf0-06a6-4188-bee7-78a1c83374dd","Type":"ContainerStarted","Data":"0a200cc565189229ec529ab5804076832b7fcc3c71cd1945e8edf7507a619613"} Nov 28 14:49:31 crc kubenswrapper[4817]: I1128 14:49:31.373517 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45dc8bf0-06a6-4188-bee7-78a1c83374dd","Type":"ContainerStarted","Data":"ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b"} Nov 28 14:49:31 crc kubenswrapper[4817]: I1128 14:49:31.373922 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45dc8bf0-06a6-4188-bee7-78a1c83374dd","Type":"ContainerStarted","Data":"8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753"} Nov 28 14:49:31 crc kubenswrapper[4817]: I1128 14:49:31.393710 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.393686413 podStartE2EDuration="2.393686413s" podCreationTimestamp="2025-11-28 14:49:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:31.388812602 +0000 UTC m=+1193.976790868" watchObservedRunningTime="2025-11-28 14:49:31.393686413 +0000 UTC m=+1193.981664689" Nov 28 14:49:33 crc kubenswrapper[4817]: I1128 14:49:33.009838 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 14:49:33 crc kubenswrapper[4817]: I1128 14:49:33.653333 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 14:49:33 crc kubenswrapper[4817]: I1128 14:49:33.653382 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 14:49:34 crc kubenswrapper[4817]: I1128 14:49:34.669914 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 14:49:34 crc kubenswrapper[4817]: I1128 14:49:34.669955 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 14:49:35 crc kubenswrapper[4817]: I1128 14:49:35.176863 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 14:49:36 crc kubenswrapper[4817]: I1128 14:49:36.715305 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 28 14:49:38 crc kubenswrapper[4817]: I1128 14:49:38.009330 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 14:49:38 crc kubenswrapper[4817]: I1128 14:49:38.039558 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 14:49:38 crc kubenswrapper[4817]: I1128 14:49:38.472141 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 14:49:38 crc kubenswrapper[4817]: I1128 14:49:38.748134 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:49:38 crc kubenswrapper[4817]: I1128 14:49:38.748606 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="bef81fb8-da0d-4735-8882-5fc41530730c" containerName="kube-state-metrics" containerID="cri-o://7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34" gracePeriod=30 Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.324114 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.447622 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7jxl\" (UniqueName: \"kubernetes.io/projected/bef81fb8-da0d-4735-8882-5fc41530730c-kube-api-access-m7jxl\") pod \"bef81fb8-da0d-4735-8882-5fc41530730c\" (UID: \"bef81fb8-da0d-4735-8882-5fc41530730c\") " Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.452865 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef81fb8-da0d-4735-8882-5fc41530730c-kube-api-access-m7jxl" (OuterVolumeSpecName: "kube-api-access-m7jxl") pod "bef81fb8-da0d-4735-8882-5fc41530730c" (UID: "bef81fb8-da0d-4735-8882-5fc41530730c"). InnerVolumeSpecName "kube-api-access-m7jxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.460235 4817 generic.go:334] "Generic (PLEG): container finished" podID="bef81fb8-da0d-4735-8882-5fc41530730c" containerID="7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34" exitCode=2 Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.460286 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.460358 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bef81fb8-da0d-4735-8882-5fc41530730c","Type":"ContainerDied","Data":"7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34"} Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.460388 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bef81fb8-da0d-4735-8882-5fc41530730c","Type":"ContainerDied","Data":"8aa73995084e5522af7efdd0429259d5167e3bc809b9c8d546228ced98debf08"} Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.460408 4817 scope.go:117] "RemoveContainer" containerID="7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.515532 4817 scope.go:117] "RemoveContainer" containerID="7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.515677 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:49:39 crc kubenswrapper[4817]: E1128 14:49:39.516027 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34\": container with ID starting with 7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34 not found: ID does not exist" containerID="7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.516063 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34"} err="failed to get container status \"7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34\": rpc error: code = NotFound desc = could not find container \"7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34\": container with ID starting with 7a3e49d8bfa753873508b90bb89e5d6f67f95baae83aaab74aad5ea22044fa34 not found: ID does not exist" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.533824 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.544793 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:49:39 crc kubenswrapper[4817]: E1128 14:49:39.545252 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef81fb8-da0d-4735-8882-5fc41530730c" containerName="kube-state-metrics" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.545268 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef81fb8-da0d-4735-8882-5fc41530730c" containerName="kube-state-metrics" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.545465 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef81fb8-da0d-4735-8882-5fc41530730c" containerName="kube-state-metrics" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.546133 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.550125 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7jxl\" (UniqueName: \"kubernetes.io/projected/bef81fb8-da0d-4735-8882-5fc41530730c-kube-api-access-m7jxl\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.551434 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.551763 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.559384 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.652020 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.652084 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.652126 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84n6l\" (UniqueName: \"kubernetes.io/projected/95d4474e-82dd-405a-a280-726ce4f894dc-kube-api-access-84n6l\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.652341 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.752586 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef81fb8-da0d-4735-8882-5fc41530730c" path="/var/lib/kubelet/pods/bef81fb8-da0d-4735-8882-5fc41530730c/volumes" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.753154 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.753184 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.754261 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.754339 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.754378 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84n6l\" (UniqueName: \"kubernetes.io/projected/95d4474e-82dd-405a-a280-726ce4f894dc-kube-api-access-84n6l\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.754439 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.759045 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.760576 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.760873 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d4474e-82dd-405a-a280-726ce4f894dc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.776419 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84n6l\" (UniqueName: \"kubernetes.io/projected/95d4474e-82dd-405a-a280-726ce4f894dc-kube-api-access-84n6l\") pod \"kube-state-metrics-0\" (UID: \"95d4474e-82dd-405a-a280-726ce4f894dc\") " pod="openstack/kube-state-metrics-0" Nov 28 14:49:39 crc kubenswrapper[4817]: I1128 14:49:39.866457 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 14:49:40 crc kubenswrapper[4817]: W1128 14:49:40.303499 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95d4474e_82dd_405a_a280_726ce4f894dc.slice/crio-15f2c524986869b14612d2111e4eba80c0e9d57d08e0abf8a6b7d6de22bd91be WatchSource:0}: Error finding container 15f2c524986869b14612d2111e4eba80c0e9d57d08e0abf8a6b7d6de22bd91be: Status 404 returned error can't find the container with id 15f2c524986869b14612d2111e4eba80c0e9d57d08e0abf8a6b7d6de22bd91be Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.307627 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.463621 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.463908 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-central-agent" containerID="cri-o://a176b6d644c4b1253ca8fd7b48b44f3f4bb2668a409ce0bd25f39f2212e196f1" gracePeriod=30 Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.463979 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="proxy-httpd" containerID="cri-o://c2d3804c26ea1e09e954d8877f509b0ce97f55b79c2ac5710c531c1f5d495fe3" gracePeriod=30 Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.463979 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="sg-core" containerID="cri-o://e565165da831a8f924eb15a51d3d281e0465304876aafe359039d345234bdcf9" gracePeriod=30 Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.464022 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-notification-agent" containerID="cri-o://33e8a4a7124d1c0e454b1c76980a764a4580f33aff0136fc6bbcf32a32042d66" gracePeriod=30 Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.475401 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95d4474e-82dd-405a-a280-726ce4f894dc","Type":"ContainerStarted","Data":"15f2c524986869b14612d2111e4eba80c0e9d57d08e0abf8a6b7d6de22bd91be"} Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.822943 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 14:49:40 crc kubenswrapper[4817]: I1128 14:49:40.822948 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.488426 4817 generic.go:334] "Generic (PLEG): container finished" podID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerID="c2d3804c26ea1e09e954d8877f509b0ce97f55b79c2ac5710c531c1f5d495fe3" exitCode=0 Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.488482 4817 generic.go:334] "Generic (PLEG): container finished" podID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerID="e565165da831a8f924eb15a51d3d281e0465304876aafe359039d345234bdcf9" exitCode=2 Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.488485 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerDied","Data":"c2d3804c26ea1e09e954d8877f509b0ce97f55b79c2ac5710c531c1f5d495fe3"} Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.488509 4817 generic.go:334] "Generic (PLEG): container finished" podID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerID="a176b6d644c4b1253ca8fd7b48b44f3f4bb2668a409ce0bd25f39f2212e196f1" exitCode=0 Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.488532 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerDied","Data":"e565165da831a8f924eb15a51d3d281e0465304876aafe359039d345234bdcf9"} Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.488543 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerDied","Data":"a176b6d644c4b1253ca8fd7b48b44f3f4bb2668a409ce0bd25f39f2212e196f1"} Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.491013 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95d4474e-82dd-405a-a280-726ce4f894dc","Type":"ContainerStarted","Data":"913915c13958a7f9750f19b65e6b0ae295be0808ceece19466e9cda70de3567e"} Nov 28 14:49:41 crc kubenswrapper[4817]: I1128 14:49:41.491300 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.513919 4817 generic.go:334] "Generic (PLEG): container finished" podID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerID="33e8a4a7124d1c0e454b1c76980a764a4580f33aff0136fc6bbcf32a32042d66" exitCode=0 Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.514028 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerDied","Data":"33e8a4a7124d1c0e454b1c76980a764a4580f33aff0136fc6bbcf32a32042d66"} Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.675868 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.699380 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.310233638 podStartE2EDuration="3.699362756s" podCreationTimestamp="2025-11-28 14:49:39 +0000 UTC" firstStartedPulling="2025-11-28 14:49:40.305007706 +0000 UTC m=+1202.892986002" lastFinishedPulling="2025-11-28 14:49:40.694136864 +0000 UTC m=+1203.282115120" observedRunningTime="2025-11-28 14:49:41.514062867 +0000 UTC m=+1204.102041133" watchObservedRunningTime="2025-11-28 14:49:42.699362756 +0000 UTC m=+1205.287341022" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811121 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-run-httpd\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811445 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-sg-core-conf-yaml\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811470 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-combined-ca-bundle\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811505 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-config-data\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811552 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-scripts\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811590 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c7pt\" (UniqueName: \"kubernetes.io/projected/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-kube-api-access-6c7pt\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811612 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.811731 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-log-httpd\") pod \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\" (UID: \"58cea07d-f2c6-4576-8e86-d28b9de0bdd0\") " Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.812141 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.812378 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.817373 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-scripts" (OuterVolumeSpecName: "scripts") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.817559 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-kube-api-access-6c7pt" (OuterVolumeSpecName: "kube-api-access-6c7pt") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "kube-api-access-6c7pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.839955 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.888578 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913526 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-config-data" (OuterVolumeSpecName: "config-data") pod "58cea07d-f2c6-4576-8e86-d28b9de0bdd0" (UID: "58cea07d-f2c6-4576-8e86-d28b9de0bdd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913783 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913822 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c7pt\" (UniqueName: \"kubernetes.io/projected/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-kube-api-access-6c7pt\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913835 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913845 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913853 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:42 crc kubenswrapper[4817]: I1128 14:49:42.913860 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58cea07d-f2c6-4576-8e86-d28b9de0bdd0-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.525596 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58cea07d-f2c6-4576-8e86-d28b9de0bdd0","Type":"ContainerDied","Data":"d3e84e0038ee2e7e50f3a02bf1183ebc5d60026a0eed418b78d33d4fbb8aa3f4"} Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.525674 4817 scope.go:117] "RemoveContainer" containerID="c2d3804c26ea1e09e954d8877f509b0ce97f55b79c2ac5710c531c1f5d495fe3" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.525843 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.564316 4817 scope.go:117] "RemoveContainer" containerID="e565165da831a8f924eb15a51d3d281e0465304876aafe359039d345234bdcf9" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.568549 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.586485 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.590114 4817 scope.go:117] "RemoveContainer" containerID="33e8a4a7124d1c0e454b1c76980a764a4580f33aff0136fc6bbcf32a32042d66" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605134 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:43 crc kubenswrapper[4817]: E1128 14:49:43.605600 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-central-agent" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605624 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-central-agent" Nov 28 14:49:43 crc kubenswrapper[4817]: E1128 14:49:43.605633 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="sg-core" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605639 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="sg-core" Nov 28 14:49:43 crc kubenswrapper[4817]: E1128 14:49:43.605650 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-notification-agent" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605657 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-notification-agent" Nov 28 14:49:43 crc kubenswrapper[4817]: E1128 14:49:43.605686 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="proxy-httpd" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605693 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="proxy-httpd" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605914 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-notification-agent" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605939 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="ceilometer-central-agent" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605955 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="sg-core" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.605975 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" containerName="proxy-httpd" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.607962 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.610779 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.610970 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.611870 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.622377 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.623773 4817 scope.go:117] "RemoveContainer" containerID="a176b6d644c4b1253ca8fd7b48b44f3f4bb2668a409ce0bd25f39f2212e196f1" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.660622 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.665659 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.672199 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.727854 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.727938 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.727983 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-config-data\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.728020 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-log-httpd\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.728075 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-scripts\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.728123 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-run-httpd\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.728162 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.728229 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnjm2\" (UniqueName: \"kubernetes.io/projected/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-kube-api-access-qnjm2\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.754312 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58cea07d-f2c6-4576-8e86-d28b9de0bdd0" path="/var/lib/kubelet/pods/58cea07d-f2c6-4576-8e86-d28b9de0bdd0/volumes" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.829951 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-run-httpd\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830348 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830418 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnjm2\" (UniqueName: \"kubernetes.io/projected/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-kube-api-access-qnjm2\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830485 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-run-httpd\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830558 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830620 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830647 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-config-data\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830674 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-log-httpd\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.830712 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-scripts\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.831463 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-log-httpd\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.835882 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.836826 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.837780 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-config-data\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.837887 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-scripts\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.838414 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.856398 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnjm2\" (UniqueName: \"kubernetes.io/projected/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-kube-api-access-qnjm2\") pod \"ceilometer-0\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " pod="openstack/ceilometer-0" Nov 28 14:49:43 crc kubenswrapper[4817]: I1128 14:49:43.928132 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:44 crc kubenswrapper[4817]: I1128 14:49:44.428566 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:44 crc kubenswrapper[4817]: I1128 14:49:44.542758 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerStarted","Data":"8c24879084311019145da93360324b9974ac09a776f7976426d8024b95d48eac"} Nov 28 14:49:44 crc kubenswrapper[4817]: I1128 14:49:44.549562 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 14:49:45 crc kubenswrapper[4817]: I1128 14:49:45.553858 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerStarted","Data":"bd2f6fc49801ee34e714e7e9806520526c076e958179c1b894246cf244b57e89"} Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.482232 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.565617 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerStarted","Data":"89d34d5446c3447fa3de9bb3a85d87d29e9c46362b18c9213f697a661ac01a6c"} Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.569279 4817 generic.go:334] "Generic (PLEG): container finished" podID="1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" containerID="1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66" exitCode=137 Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.569603 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.570503 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af","Type":"ContainerDied","Data":"1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66"} Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.570548 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af","Type":"ContainerDied","Data":"9b4dfb4ba412a8cdfcf21623f9f57e1d26816429ba57922d6ac74f91c78dbe34"} Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.570571 4817 scope.go:117] "RemoveContainer" containerID="1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.599264 4817 scope.go:117] "RemoveContainer" containerID="1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66" Nov 28 14:49:46 crc kubenswrapper[4817]: E1128 14:49:46.600376 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66\": container with ID starting with 1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66 not found: ID does not exist" containerID="1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.600435 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66"} err="failed to get container status \"1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66\": rpc error: code = NotFound desc = could not find container \"1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66\": container with ID starting with 1047e282746924c7cceff853f5a347c7834395a8fcaf822c6e4e9be537d0fa66 not found: ID does not exist" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.615427 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-combined-ca-bundle\") pod \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.615513 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwkh4\" (UniqueName: \"kubernetes.io/projected/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-kube-api-access-kwkh4\") pod \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.615642 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-config-data\") pod \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\" (UID: \"1c4b2b99-b7d9-47b9-8b29-c61eb28d17af\") " Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.624127 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-kube-api-access-kwkh4" (OuterVolumeSpecName: "kube-api-access-kwkh4") pod "1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" (UID: "1c4b2b99-b7d9-47b9-8b29-c61eb28d17af"). InnerVolumeSpecName "kube-api-access-kwkh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.646154 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-config-data" (OuterVolumeSpecName: "config-data") pod "1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" (UID: "1c4b2b99-b7d9-47b9-8b29-c61eb28d17af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.657912 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" (UID: "1c4b2b99-b7d9-47b9-8b29-c61eb28d17af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.718081 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwkh4\" (UniqueName: \"kubernetes.io/projected/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-kube-api-access-kwkh4\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.718119 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.718133 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:46 crc kubenswrapper[4817]: I1128 14:49:46.997207 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.005952 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.018795 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:47 crc kubenswrapper[4817]: E1128 14:49:47.019234 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.019253 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.019523 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.020695 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.027558 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.027609 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.030000 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.031881 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.124153 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.124224 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scnxf\" (UniqueName: \"kubernetes.io/projected/514053b0-d172-46fb-ac7c-26bd118d83f0-kube-api-access-scnxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.124324 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.124371 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.124478 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.226882 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.227166 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.227607 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.228650 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.228789 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scnxf\" (UniqueName: \"kubernetes.io/projected/514053b0-d172-46fb-ac7c-26bd118d83f0-kube-api-access-scnxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.234360 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.236679 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.237295 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.240801 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/514053b0-d172-46fb-ac7c-26bd118d83f0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.247488 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scnxf\" (UniqueName: \"kubernetes.io/projected/514053b0-d172-46fb-ac7c-26bd118d83f0-kube-api-access-scnxf\") pod \"nova-cell1-novncproxy-0\" (UID: \"514053b0-d172-46fb-ac7c-26bd118d83f0\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.342179 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.583452 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerStarted","Data":"8f6c3e18fc529a5e54215b9f1fa9ace1e881a7f756bf23ca99a83a149a1340ac"} Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.750013 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c4b2b99-b7d9-47b9-8b29-c61eb28d17af" path="/var/lib/kubelet/pods/1c4b2b99-b7d9-47b9-8b29-c61eb28d17af/volumes" Nov 28 14:49:47 crc kubenswrapper[4817]: I1128 14:49:47.838842 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 14:49:47 crc kubenswrapper[4817]: W1128 14:49:47.846466 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod514053b0_d172_46fb_ac7c_26bd118d83f0.slice/crio-69fd1337c572ce191ef6ed366e34fed7750b68d8d50aad5b5eee1b5d78119301 WatchSource:0}: Error finding container 69fd1337c572ce191ef6ed366e34fed7750b68d8d50aad5b5eee1b5d78119301: Status 404 returned error can't find the container with id 69fd1337c572ce191ef6ed366e34fed7750b68d8d50aad5b5eee1b5d78119301 Nov 28 14:49:48 crc kubenswrapper[4817]: I1128 14:49:48.599846 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"514053b0-d172-46fb-ac7c-26bd118d83f0","Type":"ContainerStarted","Data":"b5d23feb9cd8db4ed69eab2092e26fee7b1814007ae582373dd3db2cdddeae72"} Nov 28 14:49:48 crc kubenswrapper[4817]: I1128 14:49:48.600512 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"514053b0-d172-46fb-ac7c-26bd118d83f0","Type":"ContainerStarted","Data":"69fd1337c572ce191ef6ed366e34fed7750b68d8d50aad5b5eee1b5d78119301"} Nov 28 14:49:48 crc kubenswrapper[4817]: I1128 14:49:48.609347 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerStarted","Data":"d57743f5c53c3ca5026a6731383404e67c1cfa5b383c030b8cadb0ceb66f6831"} Nov 28 14:49:48 crc kubenswrapper[4817]: I1128 14:49:48.610785 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:49:48 crc kubenswrapper[4817]: I1128 14:49:48.671238 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.7685371380000001 podStartE2EDuration="5.671196535s" podCreationTimestamp="2025-11-28 14:49:43 +0000 UTC" firstStartedPulling="2025-11-28 14:49:44.427160346 +0000 UTC m=+1207.015138612" lastFinishedPulling="2025-11-28 14:49:48.329819743 +0000 UTC m=+1210.917798009" observedRunningTime="2025-11-28 14:49:48.668353684 +0000 UTC m=+1211.256331970" watchObservedRunningTime="2025-11-28 14:49:48.671196535 +0000 UTC m=+1211.259174811" Nov 28 14:49:48 crc kubenswrapper[4817]: I1128 14:49:48.676238 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.67622787 podStartE2EDuration="2.67622787s" podCreationTimestamp="2025-11-28 14:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:48.632326399 +0000 UTC m=+1211.220304695" watchObservedRunningTime="2025-11-28 14:49:48.67622787 +0000 UTC m=+1211.264206156" Nov 28 14:49:49 crc kubenswrapper[4817]: I1128 14:49:49.750112 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 14:49:49 crc kubenswrapper[4817]: I1128 14:49:49.751060 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 14:49:49 crc kubenswrapper[4817]: I1128 14:49:49.756443 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 14:49:49 crc kubenswrapper[4817]: I1128 14:49:49.762315 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 14:49:49 crc kubenswrapper[4817]: I1128 14:49:49.886304 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.637867 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.645881 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.827785 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-f95c7"] Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.829629 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.849776 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-f95c7"] Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.903765 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.903852 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.903891 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.903943 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stcnk\" (UniqueName: \"kubernetes.io/projected/a1097a83-dca3-4ad8-9793-8a68affb77da-kube-api-access-stcnk\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.903959 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:50 crc kubenswrapper[4817]: I1128 14:49:50.904010 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-config\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.006165 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stcnk\" (UniqueName: \"kubernetes.io/projected/a1097a83-dca3-4ad8-9793-8a68affb77da-kube-api-access-stcnk\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.006223 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.006309 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-config\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.006396 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.006437 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.006460 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.007508 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-config\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.007559 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.007580 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.008048 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.011413 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.027438 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stcnk\" (UniqueName: \"kubernetes.io/projected/a1097a83-dca3-4ad8-9793-8a68affb77da-kube-api-access-stcnk\") pod \"dnsmasq-dns-6b7bbf7cf9-f95c7\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.173282 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:51 crc kubenswrapper[4817]: W1128 14:49:51.676712 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1097a83_dca3_4ad8_9793_8a68affb77da.slice/crio-8e1c6a71cc575ab8a014d301a176c3a50f2e3c0f2c3b938f8d4fc2c2a732d32e WatchSource:0}: Error finding container 8e1c6a71cc575ab8a014d301a176c3a50f2e3c0f2c3b938f8d4fc2c2a732d32e: Status 404 returned error can't find the container with id 8e1c6a71cc575ab8a014d301a176c3a50f2e3c0f2c3b938f8d4fc2c2a732d32e Nov 28 14:49:51 crc kubenswrapper[4817]: I1128 14:49:51.678632 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-f95c7"] Nov 28 14:49:52 crc kubenswrapper[4817]: I1128 14:49:52.342648 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:52 crc kubenswrapper[4817]: I1128 14:49:52.652100 4817 generic.go:334] "Generic (PLEG): container finished" podID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerID="5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e" exitCode=0 Nov 28 14:49:52 crc kubenswrapper[4817]: I1128 14:49:52.653018 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" event={"ID":"a1097a83-dca3-4ad8-9793-8a68affb77da","Type":"ContainerDied","Data":"5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e"} Nov 28 14:49:52 crc kubenswrapper[4817]: I1128 14:49:52.653055 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" event={"ID":"a1097a83-dca3-4ad8-9793-8a68affb77da","Type":"ContainerStarted","Data":"8e1c6a71cc575ab8a014d301a176c3a50f2e3c0f2c3b938f8d4fc2c2a732d32e"} Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.363974 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.663911 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" event={"ID":"a1097a83-dca3-4ad8-9793-8a68affb77da","Type":"ContainerStarted","Data":"5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778"} Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.664039 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-log" containerID="cri-o://8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753" gracePeriod=30 Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.664119 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-api" containerID="cri-o://ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b" gracePeriod=30 Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.664463 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.691463 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" podStartSLOduration=3.691442078 podStartE2EDuration="3.691442078s" podCreationTimestamp="2025-11-28 14:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:53.687195593 +0000 UTC m=+1216.275173869" watchObservedRunningTime="2025-11-28 14:49:53.691442078 +0000 UTC m=+1216.279420344" Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.960157 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.961883 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-central-agent" containerID="cri-o://bd2f6fc49801ee34e714e7e9806520526c076e958179c1b894246cf244b57e89" gracePeriod=30 Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.962054 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="proxy-httpd" containerID="cri-o://d57743f5c53c3ca5026a6731383404e67c1cfa5b383c030b8cadb0ceb66f6831" gracePeriod=30 Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.962079 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-notification-agent" containerID="cri-o://89d34d5446c3447fa3de9bb3a85d87d29e9c46362b18c9213f697a661ac01a6c" gracePeriod=30 Nov 28 14:49:53 crc kubenswrapper[4817]: I1128 14:49:53.962142 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="sg-core" containerID="cri-o://8f6c3e18fc529a5e54215b9f1fa9ace1e881a7f756bf23ca99a83a149a1340ac" gracePeriod=30 Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.045813 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.046128 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.687741 4817 generic.go:334] "Generic (PLEG): container finished" podID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerID="d57743f5c53c3ca5026a6731383404e67c1cfa5b383c030b8cadb0ceb66f6831" exitCode=0 Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688042 4817 generic.go:334] "Generic (PLEG): container finished" podID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerID="8f6c3e18fc529a5e54215b9f1fa9ace1e881a7f756bf23ca99a83a149a1340ac" exitCode=2 Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688050 4817 generic.go:334] "Generic (PLEG): container finished" podID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerID="89d34d5446c3447fa3de9bb3a85d87d29e9c46362b18c9213f697a661ac01a6c" exitCode=0 Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688057 4817 generic.go:334] "Generic (PLEG): container finished" podID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerID="bd2f6fc49801ee34e714e7e9806520526c076e958179c1b894246cf244b57e89" exitCode=0 Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688100 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerDied","Data":"d57743f5c53c3ca5026a6731383404e67c1cfa5b383c030b8cadb0ceb66f6831"} Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688125 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerDied","Data":"8f6c3e18fc529a5e54215b9f1fa9ace1e881a7f756bf23ca99a83a149a1340ac"} Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688134 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerDied","Data":"89d34d5446c3447fa3de9bb3a85d87d29e9c46362b18c9213f697a661ac01a6c"} Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.688142 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerDied","Data":"bd2f6fc49801ee34e714e7e9806520526c076e958179c1b894246cf244b57e89"} Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.693549 4817 generic.go:334] "Generic (PLEG): container finished" podID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerID="8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753" exitCode=143 Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.694277 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45dc8bf0-06a6-4188-bee7-78a1c83374dd","Type":"ContainerDied","Data":"8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753"} Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.753196 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875591 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-combined-ca-bundle\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875764 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-log-httpd\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875800 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnjm2\" (UniqueName: \"kubernetes.io/projected/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-kube-api-access-qnjm2\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875845 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-config-data\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875862 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-ceilometer-tls-certs\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875916 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-sg-core-conf-yaml\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875942 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-scripts\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.875976 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-run-httpd\") pod \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\" (UID: \"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae\") " Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.876417 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.876953 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.880992 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-scripts" (OuterVolumeSpecName: "scripts") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.883936 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-kube-api-access-qnjm2" (OuterVolumeSpecName: "kube-api-access-qnjm2") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "kube-api-access-qnjm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.912441 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.944466 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.967341 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978502 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978529 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978539 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978550 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnjm2\" (UniqueName: \"kubernetes.io/projected/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-kube-api-access-qnjm2\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978559 4817 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978567 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:54 crc kubenswrapper[4817]: I1128 14:49:54.978575 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.013440 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-config-data" (OuterVolumeSpecName: "config-data") pod "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" (UID: "826cfc9f-cdb1-4d03-984b-17d7f3dd46ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.079945 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.707248 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"826cfc9f-cdb1-4d03-984b-17d7f3dd46ae","Type":"ContainerDied","Data":"8c24879084311019145da93360324b9974ac09a776f7976426d8024b95d48eac"} Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.707297 4817 scope.go:117] "RemoveContainer" containerID="d57743f5c53c3ca5026a6731383404e67c1cfa5b383c030b8cadb0ceb66f6831" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.707395 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.734083 4817 scope.go:117] "RemoveContainer" containerID="8f6c3e18fc529a5e54215b9f1fa9ace1e881a7f756bf23ca99a83a149a1340ac" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.757310 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.757357 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.765681 4817 scope.go:117] "RemoveContainer" containerID="89d34d5446c3447fa3de9bb3a85d87d29e9c46362b18c9213f697a661ac01a6c" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.779845 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:55 crc kubenswrapper[4817]: E1128 14:49:55.781076 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="proxy-httpd" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781097 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="proxy-httpd" Nov 28 14:49:55 crc kubenswrapper[4817]: E1128 14:49:55.781117 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-central-agent" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781147 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-central-agent" Nov 28 14:49:55 crc kubenswrapper[4817]: E1128 14:49:55.781159 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="sg-core" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781165 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="sg-core" Nov 28 14:49:55 crc kubenswrapper[4817]: E1128 14:49:55.781186 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-notification-agent" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781191 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-notification-agent" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781391 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="sg-core" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781403 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-notification-agent" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781422 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="ceilometer-central-agent" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.781438 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" containerName="proxy-httpd" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.783052 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.786460 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.786632 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.786807 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.798406 4817 scope.go:117] "RemoveContainer" containerID="bd2f6fc49801ee34e714e7e9806520526c076e958179c1b894246cf244b57e89" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.810921 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.895689 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-config-data\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.895874 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.895974 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpjn6\" (UniqueName: \"kubernetes.io/projected/c589da35-8233-46bc-a0da-4c00355281be-kube-api-access-lpjn6\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.896019 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.896161 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.896220 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-log-httpd\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.896270 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-scripts\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.896364 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-run-httpd\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998489 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpjn6\" (UniqueName: \"kubernetes.io/projected/c589da35-8233-46bc-a0da-4c00355281be-kube-api-access-lpjn6\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998547 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998605 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998624 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-log-httpd\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998644 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-scripts\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998695 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-run-httpd\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998712 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-config-data\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.998772 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:55 crc kubenswrapper[4817]: I1128 14:49:55.999689 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-log-httpd\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.002022 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-run-httpd\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.003957 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.004076 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-config-data\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.010903 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.012034 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-scripts\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.014411 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.017600 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpjn6\" (UniqueName: \"kubernetes.io/projected/c589da35-8233-46bc-a0da-4c00355281be-kube-api-access-lpjn6\") pod \"ceilometer-0\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.099564 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.544165 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 14:49:56 crc kubenswrapper[4817]: I1128 14:49:56.722984 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerStarted","Data":"cd35d0ed3d63ed8e55ae788658af0eb7c4305ec67e6c620c37999123ab12cf2b"} Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.342640 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.360602 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.361377 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.428890 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dc8bf0-06a6-4188-bee7-78a1c83374dd-logs\") pod \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.428958 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-combined-ca-bundle\") pod \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.429029 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-config-data\") pod \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.429237 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m42sg\" (UniqueName: \"kubernetes.io/projected/45dc8bf0-06a6-4188-bee7-78a1c83374dd-kube-api-access-m42sg\") pod \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\" (UID: \"45dc8bf0-06a6-4188-bee7-78a1c83374dd\") " Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.430406 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45dc8bf0-06a6-4188-bee7-78a1c83374dd-logs" (OuterVolumeSpecName: "logs") pod "45dc8bf0-06a6-4188-bee7-78a1c83374dd" (UID: "45dc8bf0-06a6-4188-bee7-78a1c83374dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.437125 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45dc8bf0-06a6-4188-bee7-78a1c83374dd-kube-api-access-m42sg" (OuterVolumeSpecName: "kube-api-access-m42sg") pod "45dc8bf0-06a6-4188-bee7-78a1c83374dd" (UID: "45dc8bf0-06a6-4188-bee7-78a1c83374dd"). InnerVolumeSpecName "kube-api-access-m42sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.545932 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m42sg\" (UniqueName: \"kubernetes.io/projected/45dc8bf0-06a6-4188-bee7-78a1c83374dd-kube-api-access-m42sg\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.545966 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dc8bf0-06a6-4188-bee7-78a1c83374dd-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.558785 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-config-data" (OuterVolumeSpecName: "config-data") pod "45dc8bf0-06a6-4188-bee7-78a1c83374dd" (UID: "45dc8bf0-06a6-4188-bee7-78a1c83374dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.577697 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45dc8bf0-06a6-4188-bee7-78a1c83374dd" (UID: "45dc8bf0-06a6-4188-bee7-78a1c83374dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.648219 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.648266 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dc8bf0-06a6-4188-bee7-78a1c83374dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.738883 4817 generic.go:334] "Generic (PLEG): container finished" podID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerID="ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b" exitCode=0 Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.744686 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.756193 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826cfc9f-cdb1-4d03-984b-17d7f3dd46ae" path="/var/lib/kubelet/pods/826cfc9f-cdb1-4d03-984b-17d7f3dd46ae/volumes" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.761181 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45dc8bf0-06a6-4188-bee7-78a1c83374dd","Type":"ContainerDied","Data":"ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b"} Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.761232 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45dc8bf0-06a6-4188-bee7-78a1c83374dd","Type":"ContainerDied","Data":"0a200cc565189229ec529ab5804076832b7fcc3c71cd1945e8edf7507a619613"} Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.761259 4817 scope.go:117] "RemoveContainer" containerID="ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.776669 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.809319 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.821349 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.832592 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:57 crc kubenswrapper[4817]: E1128 14:49:57.833039 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-api" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.833054 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-api" Nov 28 14:49:57 crc kubenswrapper[4817]: E1128 14:49:57.833079 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-log" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.833085 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-log" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.833283 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-log" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.833301 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" containerName="nova-api-api" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.834923 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.836980 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.837205 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.837517 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.869752 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.955624 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f891811-9a93-4453-85c3-1ff1f801caeb-logs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.955959 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fqks\" (UniqueName: \"kubernetes.io/projected/9f891811-9a93-4453-85c3-1ff1f801caeb-kube-api-access-4fqks\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.955993 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.956079 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.956111 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-config-data\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.956140 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-public-tls-certs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.969855 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dfjsl"] Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.971571 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.973661 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.973947 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 28 14:49:57 crc kubenswrapper[4817]: I1128 14:49:57.982283 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dfjsl"] Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.057700 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f891811-9a93-4453-85c3-1ff1f801caeb-logs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.057806 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fqks\" (UniqueName: \"kubernetes.io/projected/9f891811-9a93-4453-85c3-1ff1f801caeb-kube-api-access-4fqks\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.057842 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8fks\" (UniqueName: \"kubernetes.io/projected/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-kube-api-access-c8fks\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.058235 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.058239 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f891811-9a93-4453-85c3-1ff1f801caeb-logs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.058357 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-config-data\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.058404 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.058526 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-config-data\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.059075 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-public-tls-certs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.059108 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.059150 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-scripts\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.063852 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.064270 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.064301 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-config-data\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.064883 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-public-tls-certs\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.080027 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fqks\" (UniqueName: \"kubernetes.io/projected/9f891811-9a93-4453-85c3-1ff1f801caeb-kube-api-access-4fqks\") pod \"nova-api-0\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.161291 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-config-data\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.161353 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.161386 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-scripts\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.161489 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8fks\" (UniqueName: \"kubernetes.io/projected/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-kube-api-access-c8fks\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.163427 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.165036 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-scripts\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.165539 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.166238 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-config-data\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.179213 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8fks\" (UniqueName: \"kubernetes.io/projected/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-kube-api-access-c8fks\") pod \"nova-cell1-cell-mapping-dfjsl\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.242777 4817 scope.go:117] "RemoveContainer" containerID="8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.288530 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.340512 4817 scope.go:117] "RemoveContainer" containerID="ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b" Nov 28 14:49:58 crc kubenswrapper[4817]: E1128 14:49:58.343131 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b\": container with ID starting with ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b not found: ID does not exist" containerID="ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.343204 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b"} err="failed to get container status \"ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b\": rpc error: code = NotFound desc = could not find container \"ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b\": container with ID starting with ab145ebd89174aa973f4478bfd439a238162ac4679c979b2744166c911cb271b not found: ID does not exist" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.343257 4817 scope.go:117] "RemoveContainer" containerID="8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753" Nov 28 14:49:58 crc kubenswrapper[4817]: E1128 14:49:58.343680 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753\": container with ID starting with 8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753 not found: ID does not exist" containerID="8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.343702 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753"} err="failed to get container status \"8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753\": rpc error: code = NotFound desc = could not find container \"8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753\": container with ID starting with 8a29b7644b8ebc6137cc5353fb1041d07ecc1131636d77d92b32383607f9c753 not found: ID does not exist" Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.753324 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerStarted","Data":"87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9"} Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.811848 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:49:58 crc kubenswrapper[4817]: W1128 14:49:58.813447 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f891811_9a93_4453_85c3_1ff1f801caeb.slice/crio-2c97c52c787506bb0cbe308ac626506d8e057f443335d0626bacf14d486ee994 WatchSource:0}: Error finding container 2c97c52c787506bb0cbe308ac626506d8e057f443335d0626bacf14d486ee994: Status 404 returned error can't find the container with id 2c97c52c787506bb0cbe308ac626506d8e057f443335d0626bacf14d486ee994 Nov 28 14:49:58 crc kubenswrapper[4817]: I1128 14:49:58.904691 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dfjsl"] Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.755680 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45dc8bf0-06a6-4188-bee7-78a1c83374dd" path="/var/lib/kubelet/pods/45dc8bf0-06a6-4188-bee7-78a1c83374dd/volumes" Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.773528 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dfjsl" event={"ID":"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06","Type":"ContainerStarted","Data":"a4d4ba8ce44f93b94b60753c95cc2c62a36748490a2bbf51e4bb2d6e7abbd5c2"} Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.773569 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dfjsl" event={"ID":"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06","Type":"ContainerStarted","Data":"8ece27a8998e3e087ea139946f4c3bee8c2a5aa64c5852c17608f548c8554dd7"} Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.778830 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f891811-9a93-4453-85c3-1ff1f801caeb","Type":"ContainerStarted","Data":"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364"} Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.778873 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f891811-9a93-4453-85c3-1ff1f801caeb","Type":"ContainerStarted","Data":"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc"} Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.778882 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f891811-9a93-4453-85c3-1ff1f801caeb","Type":"ContainerStarted","Data":"2c97c52c787506bb0cbe308ac626506d8e057f443335d0626bacf14d486ee994"} Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.781865 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerStarted","Data":"85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52"} Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.790733 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dfjsl" podStartSLOduration=2.790698982 podStartE2EDuration="2.790698982s" podCreationTimestamp="2025-11-28 14:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:59.789088242 +0000 UTC m=+1222.377066508" watchObservedRunningTime="2025-11-28 14:49:59.790698982 +0000 UTC m=+1222.378677248" Nov 28 14:49:59 crc kubenswrapper[4817]: I1128 14:49:59.823955 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.823937788 podStartE2EDuration="2.823937788s" podCreationTimestamp="2025-11-28 14:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:49:59.812070673 +0000 UTC m=+1222.400048939" watchObservedRunningTime="2025-11-28 14:49:59.823937788 +0000 UTC m=+1222.411916054" Nov 28 14:50:00 crc kubenswrapper[4817]: I1128 14:50:00.793089 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerStarted","Data":"6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99"} Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.174880 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.253439 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lkgcd"] Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.253651 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerName="dnsmasq-dns" containerID="cri-o://dc89bf855fd55526e5c99aa2850da18a9ca6372eb5bf070a2897f42a29cc90d1" gracePeriod=10 Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.804471 4817 generic.go:334] "Generic (PLEG): container finished" podID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerID="dc89bf855fd55526e5c99aa2850da18a9ca6372eb5bf070a2897f42a29cc90d1" exitCode=0 Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.804527 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" event={"ID":"b85ff85e-86bc-45c0-940e-09c62beba8e2","Type":"ContainerDied","Data":"dc89bf855fd55526e5c99aa2850da18a9ca6372eb5bf070a2897f42a29cc90d1"} Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.804754 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" event={"ID":"b85ff85e-86bc-45c0-940e-09c62beba8e2","Type":"ContainerDied","Data":"089ca6f76699531ed55c051ca44899f5cb3021e6017fced07c710549281bff53"} Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.804770 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="089ca6f76699531ed55c051ca44899f5cb3021e6017fced07c710549281bff53" Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.833905 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.948891 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-nb\") pod \"b85ff85e-86bc-45c0-940e-09c62beba8e2\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.948951 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-config\") pod \"b85ff85e-86bc-45c0-940e-09c62beba8e2\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.949005 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-sb\") pod \"b85ff85e-86bc-45c0-940e-09c62beba8e2\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.949074 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-svc\") pod \"b85ff85e-86bc-45c0-940e-09c62beba8e2\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.949109 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-swift-storage-0\") pod \"b85ff85e-86bc-45c0-940e-09c62beba8e2\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.949160 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvd6j\" (UniqueName: \"kubernetes.io/projected/b85ff85e-86bc-45c0-940e-09c62beba8e2-kube-api-access-pvd6j\") pod \"b85ff85e-86bc-45c0-940e-09c62beba8e2\" (UID: \"b85ff85e-86bc-45c0-940e-09c62beba8e2\") " Nov 28 14:50:01 crc kubenswrapper[4817]: I1128 14:50:01.958152 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85ff85e-86bc-45c0-940e-09c62beba8e2-kube-api-access-pvd6j" (OuterVolumeSpecName: "kube-api-access-pvd6j") pod "b85ff85e-86bc-45c0-940e-09c62beba8e2" (UID: "b85ff85e-86bc-45c0-940e-09c62beba8e2"). InnerVolumeSpecName "kube-api-access-pvd6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.005149 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b85ff85e-86bc-45c0-940e-09c62beba8e2" (UID: "b85ff85e-86bc-45c0-940e-09c62beba8e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.008461 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b85ff85e-86bc-45c0-940e-09c62beba8e2" (UID: "b85ff85e-86bc-45c0-940e-09c62beba8e2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.009532 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-config" (OuterVolumeSpecName: "config") pod "b85ff85e-86bc-45c0-940e-09c62beba8e2" (UID: "b85ff85e-86bc-45c0-940e-09c62beba8e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.016149 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b85ff85e-86bc-45c0-940e-09c62beba8e2" (UID: "b85ff85e-86bc-45c0-940e-09c62beba8e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.025226 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b85ff85e-86bc-45c0-940e-09c62beba8e2" (UID: "b85ff85e-86bc-45c0-940e-09c62beba8e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.051381 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.051418 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.051429 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.051437 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.051445 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85ff85e-86bc-45c0-940e-09c62beba8e2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.051455 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvd6j\" (UniqueName: \"kubernetes.io/projected/b85ff85e-86bc-45c0-940e-09c62beba8e2-kube-api-access-pvd6j\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.815281 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerStarted","Data":"6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171"} Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.815432 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.816212 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lkgcd" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.844184 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.738914303 podStartE2EDuration="7.844163339s" podCreationTimestamp="2025-11-28 14:49:55 +0000 UTC" firstStartedPulling="2025-11-28 14:49:56.546162678 +0000 UTC m=+1219.134140954" lastFinishedPulling="2025-11-28 14:50:01.651411724 +0000 UTC m=+1224.239389990" observedRunningTime="2025-11-28 14:50:02.841569865 +0000 UTC m=+1225.429548131" watchObservedRunningTime="2025-11-28 14:50:02.844163339 +0000 UTC m=+1225.432141625" Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.865510 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lkgcd"] Nov 28 14:50:02 crc kubenswrapper[4817]: I1128 14:50:02.873120 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lkgcd"] Nov 28 14:50:03 crc kubenswrapper[4817]: I1128 14:50:03.750794 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" path="/var/lib/kubelet/pods/b85ff85e-86bc-45c0-940e-09c62beba8e2/volumes" Nov 28 14:50:04 crc kubenswrapper[4817]: I1128 14:50:04.837997 4817 generic.go:334] "Generic (PLEG): container finished" podID="0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" containerID="a4d4ba8ce44f93b94b60753c95cc2c62a36748490a2bbf51e4bb2d6e7abbd5c2" exitCode=0 Nov 28 14:50:04 crc kubenswrapper[4817]: I1128 14:50:04.838033 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dfjsl" event={"ID":"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06","Type":"ContainerDied","Data":"a4d4ba8ce44f93b94b60753c95cc2c62a36748490a2bbf51e4bb2d6e7abbd5c2"} Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.236636 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.346784 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8fks\" (UniqueName: \"kubernetes.io/projected/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-kube-api-access-c8fks\") pod \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.346927 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-scripts\") pod \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.346949 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-config-data\") pod \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.346992 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-combined-ca-bundle\") pod \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\" (UID: \"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06\") " Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.352594 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-kube-api-access-c8fks" (OuterVolumeSpecName: "kube-api-access-c8fks") pod "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" (UID: "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06"). InnerVolumeSpecName "kube-api-access-c8fks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.353060 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-scripts" (OuterVolumeSpecName: "scripts") pod "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" (UID: "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.375519 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" (UID: "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.389109 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-config-data" (OuterVolumeSpecName: "config-data") pod "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" (UID: "0d23a9c3-63c8-49a7-80e2-7f6148e0bf06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.449706 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8fks\" (UniqueName: \"kubernetes.io/projected/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-kube-api-access-c8fks\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.449770 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.449785 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.449798 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.860451 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dfjsl" event={"ID":"0d23a9c3-63c8-49a7-80e2-7f6148e0bf06","Type":"ContainerDied","Data":"8ece27a8998e3e087ea139946f4c3bee8c2a5aa64c5852c17608f548c8554dd7"} Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.860702 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ece27a8998e3e087ea139946f4c3bee8c2a5aa64c5852c17608f548c8554dd7" Nov 28 14:50:06 crc kubenswrapper[4817]: I1128 14:50:06.860527 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dfjsl" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.042115 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.042383 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-log" containerID="cri-o://f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc" gracePeriod=30 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.042522 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-api" containerID="cri-o://13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364" gracePeriod=30 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.064005 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.064265 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="dd202581-e588-42a1-bd75-7ea6a8e03b23" containerName="nova-scheduler-scheduler" containerID="cri-o://167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420" gracePeriod=30 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.081020 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.081289 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-log" containerID="cri-o://15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d" gracePeriod=30 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.081368 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-metadata" containerID="cri-o://5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0" gracePeriod=30 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.654200 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.776248 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-config-data\") pod \"9f891811-9a93-4453-85c3-1ff1f801caeb\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.776302 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-public-tls-certs\") pod \"9f891811-9a93-4453-85c3-1ff1f801caeb\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.776336 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f891811-9a93-4453-85c3-1ff1f801caeb-logs\") pod \"9f891811-9a93-4453-85c3-1ff1f801caeb\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.776536 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fqks\" (UniqueName: \"kubernetes.io/projected/9f891811-9a93-4453-85c3-1ff1f801caeb-kube-api-access-4fqks\") pod \"9f891811-9a93-4453-85c3-1ff1f801caeb\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.776627 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-internal-tls-certs\") pod \"9f891811-9a93-4453-85c3-1ff1f801caeb\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.776698 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-combined-ca-bundle\") pod \"9f891811-9a93-4453-85c3-1ff1f801caeb\" (UID: \"9f891811-9a93-4453-85c3-1ff1f801caeb\") " Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.778864 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f891811-9a93-4453-85c3-1ff1f801caeb-logs" (OuterVolumeSpecName: "logs") pod "9f891811-9a93-4453-85c3-1ff1f801caeb" (UID: "9f891811-9a93-4453-85c3-1ff1f801caeb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.783316 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f891811-9a93-4453-85c3-1ff1f801caeb-kube-api-access-4fqks" (OuterVolumeSpecName: "kube-api-access-4fqks") pod "9f891811-9a93-4453-85c3-1ff1f801caeb" (UID: "9f891811-9a93-4453-85c3-1ff1f801caeb"). InnerVolumeSpecName "kube-api-access-4fqks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.805922 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f891811-9a93-4453-85c3-1ff1f801caeb" (UID: "9f891811-9a93-4453-85c3-1ff1f801caeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.815593 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-config-data" (OuterVolumeSpecName: "config-data") pod "9f891811-9a93-4453-85c3-1ff1f801caeb" (UID: "9f891811-9a93-4453-85c3-1ff1f801caeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.847391 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9f891811-9a93-4453-85c3-1ff1f801caeb" (UID: "9f891811-9a93-4453-85c3-1ff1f801caeb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.856291 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9f891811-9a93-4453-85c3-1ff1f801caeb" (UID: "9f891811-9a93-4453-85c3-1ff1f801caeb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873526 4817 generic.go:334] "Generic (PLEG): container finished" podID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerID="13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364" exitCode=0 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873561 4817 generic.go:334] "Generic (PLEG): container finished" podID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerID="f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc" exitCode=143 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873628 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f891811-9a93-4453-85c3-1ff1f801caeb","Type":"ContainerDied","Data":"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364"} Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873659 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f891811-9a93-4453-85c3-1ff1f801caeb","Type":"ContainerDied","Data":"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc"} Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873667 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f891811-9a93-4453-85c3-1ff1f801caeb","Type":"ContainerDied","Data":"2c97c52c787506bb0cbe308ac626506d8e057f443335d0626bacf14d486ee994"} Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873682 4817 scope.go:117] "RemoveContainer" containerID="13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.873832 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.879026 4817 generic.go:334] "Generic (PLEG): container finished" podID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerID="15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d" exitCode=143 Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.879070 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6290a311-8ee7-4864-b63f-a6a5225c0df4","Type":"ContainerDied","Data":"15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d"} Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.880294 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fqks\" (UniqueName: \"kubernetes.io/projected/9f891811-9a93-4453-85c3-1ff1f801caeb-kube-api-access-4fqks\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.880355 4817 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.880368 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.880379 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.880390 4817 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f891811-9a93-4453-85c3-1ff1f801caeb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.880428 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f891811-9a93-4453-85c3-1ff1f801caeb-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.909963 4817 scope.go:117] "RemoveContainer" containerID="f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.937846 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.949323 4817 scope.go:117] "RemoveContainer" containerID="13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364" Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.950885 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364\": container with ID starting with 13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364 not found: ID does not exist" containerID="13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.950922 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364"} err="failed to get container status \"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364\": rpc error: code = NotFound desc = could not find container \"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364\": container with ID starting with 13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364 not found: ID does not exist" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.950950 4817 scope.go:117] "RemoveContainer" containerID="f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc" Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.951336 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc\": container with ID starting with f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc not found: ID does not exist" containerID="f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.951363 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc"} err="failed to get container status \"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc\": rpc error: code = NotFound desc = could not find container \"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc\": container with ID starting with f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc not found: ID does not exist" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.951377 4817 scope.go:117] "RemoveContainer" containerID="13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.951644 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364"} err="failed to get container status \"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364\": rpc error: code = NotFound desc = could not find container \"13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364\": container with ID starting with 13e136347ef73b75e4f58a913f9a7616723518ae904227be5f8549019905b364 not found: ID does not exist" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.951688 4817 scope.go:117] "RemoveContainer" containerID="f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.958896 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.963451 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc"} err="failed to get container status \"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc\": rpc error: code = NotFound desc = could not find container \"f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc\": container with ID starting with f503e3e645f9990a946ca4e7fb06f6a3a503c5ebd2816e44729d4217e718e7cc not found: ID does not exist" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.966331 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.966810 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerName="dnsmasq-dns" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.966824 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerName="dnsmasq-dns" Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.966849 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerName="init" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.966856 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerName="init" Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.966866 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-log" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.966872 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-log" Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.966878 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" containerName="nova-manage" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.966884 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" containerName="nova-manage" Nov 28 14:50:07 crc kubenswrapper[4817]: E1128 14:50:07.966901 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-api" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.966907 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-api" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.967102 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-log" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.967114 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" containerName="nova-manage" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.967126 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85ff85e-86bc-45c0-940e-09c62beba8e2" containerName="dnsmasq-dns" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.967138 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" containerName="nova-api-api" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.968773 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.971565 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.971837 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.972190 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 14:50:07 crc kubenswrapper[4817]: I1128 14:50:07.981552 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:50:08 crc kubenswrapper[4817]: E1128 14:50:08.011870 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 14:50:08 crc kubenswrapper[4817]: E1128 14:50:08.013518 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 14:50:08 crc kubenswrapper[4817]: E1128 14:50:08.019255 4817 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 14:50:08 crc kubenswrapper[4817]: E1128 14:50:08.019294 4817 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="dd202581-e588-42a1-bd75-7ea6a8e03b23" containerName="nova-scheduler-scheduler" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.084227 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-public-tls-certs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.084284 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.084422 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb11cfb0-8f52-4361-bbce-b70613a51ca6-logs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.084447 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.084503 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75zhl\" (UniqueName: \"kubernetes.io/projected/eb11cfb0-8f52-4361-bbce-b70613a51ca6-kube-api-access-75zhl\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.084552 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-config-data\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186026 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb11cfb0-8f52-4361-bbce-b70613a51ca6-logs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186071 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186118 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75zhl\" (UniqueName: \"kubernetes.io/projected/eb11cfb0-8f52-4361-bbce-b70613a51ca6-kube-api-access-75zhl\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186161 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-config-data\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186234 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-public-tls-certs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186252 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.186590 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb11cfb0-8f52-4361-bbce-b70613a51ca6-logs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.190978 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.191945 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-config-data\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.194504 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-public-tls-certs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.194864 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb11cfb0-8f52-4361-bbce-b70613a51ca6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.204666 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75zhl\" (UniqueName: \"kubernetes.io/projected/eb11cfb0-8f52-4361-bbce-b70613a51ca6-kube-api-access-75zhl\") pod \"nova-api-0\" (UID: \"eb11cfb0-8f52-4361-bbce-b70613a51ca6\") " pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.288793 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.764170 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 14:50:08 crc kubenswrapper[4817]: I1128 14:50:08.890459 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb11cfb0-8f52-4361-bbce-b70613a51ca6","Type":"ContainerStarted","Data":"a9d914f742f41114811a6f95e2f834d8e34443ed7033165569b068ae63e95a20"} Nov 28 14:50:09 crc kubenswrapper[4817]: I1128 14:50:09.749145 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f891811-9a93-4453-85c3-1ff1f801caeb" path="/var/lib/kubelet/pods/9f891811-9a93-4453-85c3-1ff1f801caeb/volumes" Nov 28 14:50:09 crc kubenswrapper[4817]: I1128 14:50:09.905265 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb11cfb0-8f52-4361-bbce-b70613a51ca6","Type":"ContainerStarted","Data":"bd0752f6aab9741d854f647ecd80502466ef199d2676d874aea379aa1dd52ffc"} Nov 28 14:50:09 crc kubenswrapper[4817]: I1128 14:50:09.905670 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eb11cfb0-8f52-4361-bbce-b70613a51ca6","Type":"ContainerStarted","Data":"21db3f4559639c3501454c2d8409f150e574701bb663535048569bd60e8b4f3e"} Nov 28 14:50:09 crc kubenswrapper[4817]: I1128 14:50:09.960177 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.960161285 podStartE2EDuration="2.960161285s" podCreationTimestamp="2025-11-28 14:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:50:09.953283264 +0000 UTC m=+1232.541261540" watchObservedRunningTime="2025-11-28 14:50:09.960161285 +0000 UTC m=+1232.548139541" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.207225 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:52002->10.217.0.202:8775: read: connection reset by peer" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.207812 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:51990->10.217.0.202:8775: read: connection reset by peer" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.710573 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.841064 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-combined-ca-bundle\") pod \"6290a311-8ee7-4864-b63f-a6a5225c0df4\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.841169 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-nova-metadata-tls-certs\") pod \"6290a311-8ee7-4864-b63f-a6a5225c0df4\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.841292 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgxgz\" (UniqueName: \"kubernetes.io/projected/6290a311-8ee7-4864-b63f-a6a5225c0df4-kube-api-access-fgxgz\") pod \"6290a311-8ee7-4864-b63f-a6a5225c0df4\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.841311 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6290a311-8ee7-4864-b63f-a6a5225c0df4-logs\") pod \"6290a311-8ee7-4864-b63f-a6a5225c0df4\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.841376 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-config-data\") pod \"6290a311-8ee7-4864-b63f-a6a5225c0df4\" (UID: \"6290a311-8ee7-4864-b63f-a6a5225c0df4\") " Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.841954 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6290a311-8ee7-4864-b63f-a6a5225c0df4-logs" (OuterVolumeSpecName: "logs") pod "6290a311-8ee7-4864-b63f-a6a5225c0df4" (UID: "6290a311-8ee7-4864-b63f-a6a5225c0df4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.842228 4817 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6290a311-8ee7-4864-b63f-a6a5225c0df4-logs\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.850990 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6290a311-8ee7-4864-b63f-a6a5225c0df4-kube-api-access-fgxgz" (OuterVolumeSpecName: "kube-api-access-fgxgz") pod "6290a311-8ee7-4864-b63f-a6a5225c0df4" (UID: "6290a311-8ee7-4864-b63f-a6a5225c0df4"). InnerVolumeSpecName "kube-api-access-fgxgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.880880 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6290a311-8ee7-4864-b63f-a6a5225c0df4" (UID: "6290a311-8ee7-4864-b63f-a6a5225c0df4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.887739 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-config-data" (OuterVolumeSpecName: "config-data") pod "6290a311-8ee7-4864-b63f-a6a5225c0df4" (UID: "6290a311-8ee7-4864-b63f-a6a5225c0df4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.907014 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6290a311-8ee7-4864-b63f-a6a5225c0df4" (UID: "6290a311-8ee7-4864-b63f-a6a5225c0df4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.918373 4817 generic.go:334] "Generic (PLEG): container finished" podID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerID="5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0" exitCode=0 Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.918425 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.918638 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6290a311-8ee7-4864-b63f-a6a5225c0df4","Type":"ContainerDied","Data":"5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0"} Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.919243 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6290a311-8ee7-4864-b63f-a6a5225c0df4","Type":"ContainerDied","Data":"1c6067aeac80518fd0e293d695a751370636c3c149566fd2a33a13441c9d76cd"} Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.919282 4817 scope.go:117] "RemoveContainer" containerID="5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.944081 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.944121 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.944134 4817 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6290a311-8ee7-4864-b63f-a6a5225c0df4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.944144 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgxgz\" (UniqueName: \"kubernetes.io/projected/6290a311-8ee7-4864-b63f-a6a5225c0df4-kube-api-access-fgxgz\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.973128 4817 scope.go:117] "RemoveContainer" containerID="15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d" Nov 28 14:50:10 crc kubenswrapper[4817]: I1128 14:50:10.988699 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.006556 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.011986 4817 scope.go:117] "RemoveContainer" containerID="5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0" Nov 28 14:50:11 crc kubenswrapper[4817]: E1128 14:50:11.012460 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0\": container with ID starting with 5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0 not found: ID does not exist" containerID="5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.012490 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0"} err="failed to get container status \"5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0\": rpc error: code = NotFound desc = could not find container \"5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0\": container with ID starting with 5d5819582535cf5190d15c58a9de181ad6c1897997eb84d86aa6cba110ff9be0 not found: ID does not exist" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.012510 4817 scope.go:117] "RemoveContainer" containerID="15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d" Nov 28 14:50:11 crc kubenswrapper[4817]: E1128 14:50:11.012766 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d\": container with ID starting with 15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d not found: ID does not exist" containerID="15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.012787 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d"} err="failed to get container status \"15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d\": rpc error: code = NotFound desc = could not find container \"15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d\": container with ID starting with 15250ad37abbb42bc4cc6827d4d2f1acdbcd22a0cdd3898d91456fd4948dd91d not found: ID does not exist" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.018646 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:50:11 crc kubenswrapper[4817]: E1128 14:50:11.019109 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-metadata" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.019127 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-metadata" Nov 28 14:50:11 crc kubenswrapper[4817]: E1128 14:50:11.019170 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-log" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.019180 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-log" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.019397 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-metadata" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.019424 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" containerName="nova-metadata-log" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.020874 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.023588 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.025630 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.032039 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.147683 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99283929-00ab-45d4-92c9-73cbb67f205f-logs\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.148041 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6ssg\" (UniqueName: \"kubernetes.io/projected/99283929-00ab-45d4-92c9-73cbb67f205f-kube-api-access-v6ssg\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.148119 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.148350 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-config-data\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.148450 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.250304 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.250384 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-config-data\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.250416 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.250488 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99283929-00ab-45d4-92c9-73cbb67f205f-logs\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.250517 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6ssg\" (UniqueName: \"kubernetes.io/projected/99283929-00ab-45d4-92c9-73cbb67f205f-kube-api-access-v6ssg\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.251334 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99283929-00ab-45d4-92c9-73cbb67f205f-logs\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.255021 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-config-data\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.255087 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.256124 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99283929-00ab-45d4-92c9-73cbb67f205f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.266849 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6ssg\" (UniqueName: \"kubernetes.io/projected/99283929-00ab-45d4-92c9-73cbb67f205f-kube-api-access-v6ssg\") pod \"nova-metadata-0\" (UID: \"99283929-00ab-45d4-92c9-73cbb67f205f\") " pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.348096 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.749601 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6290a311-8ee7-4864-b63f-a6a5225c0df4" path="/var/lib/kubelet/pods/6290a311-8ee7-4864-b63f-a6a5225c0df4/volumes" Nov 28 14:50:11 crc kubenswrapper[4817]: W1128 14:50:11.803005 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99283929_00ab_45d4_92c9_73cbb67f205f.slice/crio-212cbbf2ed8444c85e174602b2675eca9866a31a48129ae147ea9370ec68d775 WatchSource:0}: Error finding container 212cbbf2ed8444c85e174602b2675eca9866a31a48129ae147ea9370ec68d775: Status 404 returned error can't find the container with id 212cbbf2ed8444c85e174602b2675eca9866a31a48129ae147ea9370ec68d775 Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.805190 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.940066 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99283929-00ab-45d4-92c9-73cbb67f205f","Type":"ContainerStarted","Data":"212cbbf2ed8444c85e174602b2675eca9866a31a48129ae147ea9370ec68d775"} Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.947844 4817 generic.go:334] "Generic (PLEG): container finished" podID="dd202581-e588-42a1-bd75-7ea6a8e03b23" containerID="167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420" exitCode=0 Nov 28 14:50:11 crc kubenswrapper[4817]: I1128 14:50:11.947888 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd202581-e588-42a1-bd75-7ea6a8e03b23","Type":"ContainerDied","Data":"167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420"} Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.142821 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.268681 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crb5g\" (UniqueName: \"kubernetes.io/projected/dd202581-e588-42a1-bd75-7ea6a8e03b23-kube-api-access-crb5g\") pod \"dd202581-e588-42a1-bd75-7ea6a8e03b23\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.268947 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-config-data\") pod \"dd202581-e588-42a1-bd75-7ea6a8e03b23\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.268989 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-combined-ca-bundle\") pod \"dd202581-e588-42a1-bd75-7ea6a8e03b23\" (UID: \"dd202581-e588-42a1-bd75-7ea6a8e03b23\") " Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.272883 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd202581-e588-42a1-bd75-7ea6a8e03b23-kube-api-access-crb5g" (OuterVolumeSpecName: "kube-api-access-crb5g") pod "dd202581-e588-42a1-bd75-7ea6a8e03b23" (UID: "dd202581-e588-42a1-bd75-7ea6a8e03b23"). InnerVolumeSpecName "kube-api-access-crb5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.293887 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-config-data" (OuterVolumeSpecName: "config-data") pod "dd202581-e588-42a1-bd75-7ea6a8e03b23" (UID: "dd202581-e588-42a1-bd75-7ea6a8e03b23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.299697 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd202581-e588-42a1-bd75-7ea6a8e03b23" (UID: "dd202581-e588-42a1-bd75-7ea6a8e03b23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.371498 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crb5g\" (UniqueName: \"kubernetes.io/projected/dd202581-e588-42a1-bd75-7ea6a8e03b23-kube-api-access-crb5g\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.371542 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.371557 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd202581-e588-42a1-bd75-7ea6a8e03b23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.957286 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99283929-00ab-45d4-92c9-73cbb67f205f","Type":"ContainerStarted","Data":"3a1b651d171dbf4cba6bd930101ed0d8102b8ea2d42f59630f50b4fd21dabbbe"} Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.957380 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99283929-00ab-45d4-92c9-73cbb67f205f","Type":"ContainerStarted","Data":"5d3cdc6ee24abaa7a1c01c8f4999a886515fe5b004e0ccd7786697731a1788c4"} Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.959493 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dd202581-e588-42a1-bd75-7ea6a8e03b23","Type":"ContainerDied","Data":"4b4f742a0f4d18cf9599926b45a95c70ed32a55dae2aae46e1c8a5d6c30a812c"} Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.959527 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.959563 4817 scope.go:117] "RemoveContainer" containerID="167f62ff242a663ceaacd651cc100b6f22124bc663301bd6f5cee47fe68f1420" Nov 28 14:50:12 crc kubenswrapper[4817]: I1128 14:50:12.986670 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.986649962 podStartE2EDuration="2.986649962s" podCreationTimestamp="2025-11-28 14:50:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:50:12.977949345 +0000 UTC m=+1235.565927631" watchObservedRunningTime="2025-11-28 14:50:12.986649962 +0000 UTC m=+1235.574628228" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.002093 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.014418 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.022247 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:50:13 crc kubenswrapper[4817]: E1128 14:50:13.022667 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd202581-e588-42a1-bd75-7ea6a8e03b23" containerName="nova-scheduler-scheduler" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.022686 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd202581-e588-42a1-bd75-7ea6a8e03b23" containerName="nova-scheduler-scheduler" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.022909 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd202581-e588-42a1-bd75-7ea6a8e03b23" containerName="nova-scheduler-scheduler" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.023611 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.025950 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.041801 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.083128 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcffa08-235a-4101-bb4f-8933dc9102e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.083185 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd9ph\" (UniqueName: \"kubernetes.io/projected/1fcffa08-235a-4101-bb4f-8933dc9102e3-kube-api-access-gd9ph\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.083323 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcffa08-235a-4101-bb4f-8933dc9102e3-config-data\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.185311 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcffa08-235a-4101-bb4f-8933dc9102e3-config-data\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.185405 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcffa08-235a-4101-bb4f-8933dc9102e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.185451 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd9ph\" (UniqueName: \"kubernetes.io/projected/1fcffa08-235a-4101-bb4f-8933dc9102e3-kube-api-access-gd9ph\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.189891 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcffa08-235a-4101-bb4f-8933dc9102e3-config-data\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.190659 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcffa08-235a-4101-bb4f-8933dc9102e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.205525 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd9ph\" (UniqueName: \"kubernetes.io/projected/1fcffa08-235a-4101-bb4f-8933dc9102e3-kube-api-access-gd9ph\") pod \"nova-scheduler-0\" (UID: \"1fcffa08-235a-4101-bb4f-8933dc9102e3\") " pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.379963 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.747769 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd202581-e588-42a1-bd75-7ea6a8e03b23" path="/var/lib/kubelet/pods/dd202581-e588-42a1-bd75-7ea6a8e03b23/volumes" Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.838495 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 14:50:13 crc kubenswrapper[4817]: I1128 14:50:13.970277 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1fcffa08-235a-4101-bb4f-8933dc9102e3","Type":"ContainerStarted","Data":"dde8dbc2b834009e1f60ba517a520d3f3ad6e3a4dc7ca493e0b11b050717ef0a"} Nov 28 14:50:14 crc kubenswrapper[4817]: I1128 14:50:14.983835 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1fcffa08-235a-4101-bb4f-8933dc9102e3","Type":"ContainerStarted","Data":"f5e65e94722af82432a86a7230e2016cbce9c3553913c256076cecba93a86840"} Nov 28 14:50:15 crc kubenswrapper[4817]: I1128 14:50:15.022095 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.022069824 podStartE2EDuration="3.022069824s" podCreationTimestamp="2025-11-28 14:50:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:50:15.001360039 +0000 UTC m=+1237.589338335" watchObservedRunningTime="2025-11-28 14:50:15.022069824 +0000 UTC m=+1237.610048130" Nov 28 14:50:16 crc kubenswrapper[4817]: I1128 14:50:16.348233 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 14:50:16 crc kubenswrapper[4817]: I1128 14:50:16.349807 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 14:50:18 crc kubenswrapper[4817]: I1128 14:50:18.289376 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 14:50:18 crc kubenswrapper[4817]: I1128 14:50:18.289772 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 14:50:18 crc kubenswrapper[4817]: I1128 14:50:18.380265 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 14:50:19 crc kubenswrapper[4817]: I1128 14:50:19.296092 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eb11cfb0-8f52-4361-bbce-b70613a51ca6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 14:50:19 crc kubenswrapper[4817]: I1128 14:50:19.302260 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eb11cfb0-8f52-4361-bbce-b70613a51ca6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 14:50:21 crc kubenswrapper[4817]: I1128 14:50:21.349101 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 14:50:21 crc kubenswrapper[4817]: I1128 14:50:21.349460 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 14:50:22 crc kubenswrapper[4817]: I1128 14:50:22.364351 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="99283929-00ab-45d4-92c9-73cbb67f205f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 14:50:22 crc kubenswrapper[4817]: I1128 14:50:22.364431 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="99283929-00ab-45d4-92c9-73cbb67f205f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 14:50:23 crc kubenswrapper[4817]: I1128 14:50:23.380934 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 14:50:23 crc kubenswrapper[4817]: I1128 14:50:23.411300 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 14:50:24 crc kubenswrapper[4817]: I1128 14:50:24.045307 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:50:24 crc kubenswrapper[4817]: I1128 14:50:24.045368 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:50:24 crc kubenswrapper[4817]: I1128 14:50:24.045430 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:50:24 crc kubenswrapper[4817]: I1128 14:50:24.046229 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6233c40182fb7115b99ef7443b2744ab4c0eae4c9222a791b73fbfc8281a9e58"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:50:24 crc kubenswrapper[4817]: I1128 14:50:24.046294 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://6233c40182fb7115b99ef7443b2744ab4c0eae4c9222a791b73fbfc8281a9e58" gracePeriod=600 Nov 28 14:50:24 crc kubenswrapper[4817]: I1128 14:50:24.134845 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 14:50:25 crc kubenswrapper[4817]: I1128 14:50:25.112021 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="6233c40182fb7115b99ef7443b2744ab4c0eae4c9222a791b73fbfc8281a9e58" exitCode=0 Nov 28 14:50:25 crc kubenswrapper[4817]: I1128 14:50:25.112087 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"6233c40182fb7115b99ef7443b2744ab4c0eae4c9222a791b73fbfc8281a9e58"} Nov 28 14:50:25 crc kubenswrapper[4817]: I1128 14:50:25.112437 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6"} Nov 28 14:50:25 crc kubenswrapper[4817]: I1128 14:50:25.112460 4817 scope.go:117] "RemoveContainer" containerID="f352602e89e72622a4635f595006f3c005b7dadf542d0db30e6fcfcc7b955afc" Nov 28 14:50:26 crc kubenswrapper[4817]: I1128 14:50:26.113262 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 14:50:28 crc kubenswrapper[4817]: I1128 14:50:28.298810 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 14:50:28 crc kubenswrapper[4817]: I1128 14:50:28.299430 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 14:50:28 crc kubenswrapper[4817]: I1128 14:50:28.301430 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 14:50:28 crc kubenswrapper[4817]: I1128 14:50:28.307201 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 14:50:29 crc kubenswrapper[4817]: I1128 14:50:29.163368 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 14:50:29 crc kubenswrapper[4817]: I1128 14:50:29.169581 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 14:50:31 crc kubenswrapper[4817]: I1128 14:50:31.355509 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 14:50:31 crc kubenswrapper[4817]: I1128 14:50:31.356489 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 14:50:31 crc kubenswrapper[4817]: I1128 14:50:31.364582 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 14:50:32 crc kubenswrapper[4817]: I1128 14:50:32.198504 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 14:50:40 crc kubenswrapper[4817]: I1128 14:50:40.372192 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:50:41 crc kubenswrapper[4817]: I1128 14:50:41.185794 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:50:44 crc kubenswrapper[4817]: I1128 14:50:44.442901 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="rabbitmq" containerID="cri-o://0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e" gracePeriod=604796 Nov 28 14:50:45 crc kubenswrapper[4817]: I1128 14:50:45.141661 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="rabbitmq" containerID="cri-o://ed4c3b1acfccbff43f85441f72bfebd43f8cba89beaf216be01366e988d4c343" gracePeriod=604797 Nov 28 14:50:46 crc kubenswrapper[4817]: I1128 14:50:46.530117 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Nov 28 14:50:46 crc kubenswrapper[4817]: I1128 14:50:46.834511 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.030349 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.128922 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129437 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-plugins-conf\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129474 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-tls\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129539 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63b71752-f06a-4d1a-9d80-85142bbc3610-pod-info\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129630 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-erlang-cookie\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129748 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-server-conf\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129778 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63b71752-f06a-4d1a-9d80-85142bbc3610-erlang-cookie-secret\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129865 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-config-data\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129911 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdgh2\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-kube-api-access-fdgh2\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129938 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-plugins\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.129960 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-confd\") pod \"63b71752-f06a-4d1a-9d80-85142bbc3610\" (UID: \"63b71752-f06a-4d1a-9d80-85142bbc3610\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.130453 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.130977 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.134363 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.139461 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.140041 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-kube-api-access-fdgh2" (OuterVolumeSpecName: "kube-api-access-fdgh2") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "kube-api-access-fdgh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.142029 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b71752-f06a-4d1a-9d80-85142bbc3610-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.142403 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/63b71752-f06a-4d1a-9d80-85142bbc3610-pod-info" (OuterVolumeSpecName: "pod-info") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.144859 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.198144 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-config-data" (OuterVolumeSpecName: "config-data") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231138 4817 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63b71752-f06a-4d1a-9d80-85142bbc3610-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231175 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231185 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdgh2\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-kube-api-access-fdgh2\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231194 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231214 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231223 4817 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231231 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231239 4817 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63b71752-f06a-4d1a-9d80-85142bbc3610-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.231247 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.238347 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-server-conf" (OuterVolumeSpecName: "server-conf") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.254433 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.285456 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "63b71752-f06a-4d1a-9d80-85142bbc3610" (UID: "63b71752-f06a-4d1a-9d80-85142bbc3610"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.332992 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63b71752-f06a-4d1a-9d80-85142bbc3610-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.333035 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.333045 4817 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63b71752-f06a-4d1a-9d80-85142bbc3610-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.400136 4817 generic.go:334] "Generic (PLEG): container finished" podID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerID="ed4c3b1acfccbff43f85441f72bfebd43f8cba89beaf216be01366e988d4c343" exitCode=0 Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.400274 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d340d5c2-67ed-48da-826f-e1e4c244832b","Type":"ContainerDied","Data":"ed4c3b1acfccbff43f85441f72bfebd43f8cba89beaf216be01366e988d4c343"} Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.402680 4817 generic.go:334] "Generic (PLEG): container finished" podID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerID="0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e" exitCode=0 Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.402735 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63b71752-f06a-4d1a-9d80-85142bbc3610","Type":"ContainerDied","Data":"0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e"} Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.402765 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63b71752-f06a-4d1a-9d80-85142bbc3610","Type":"ContainerDied","Data":"185a4b26ff5692655011f475f0a810ea3638e19206e8f3a960240ea602406b6b"} Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.402782 4817 scope.go:117] "RemoveContainer" containerID="0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.402844 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.444692 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.463435 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.470285 4817 scope.go:117] "RemoveContainer" containerID="290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.478181 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:50:51 crc kubenswrapper[4817]: E1128 14:50:51.478955 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="rabbitmq" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.478971 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="rabbitmq" Nov 28 14:50:51 crc kubenswrapper[4817]: E1128 14:50:51.478998 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="setup-container" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.479007 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="setup-container" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.479208 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" containerName="rabbitmq" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.480434 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.483905 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.484944 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.485002 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.485304 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.485473 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.485684 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.485698 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6vf9j" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.501262 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.528188 4817 scope.go:117] "RemoveContainer" containerID="0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e" Nov 28 14:50:51 crc kubenswrapper[4817]: E1128 14:50:51.531046 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e\": container with ID starting with 0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e not found: ID does not exist" containerID="0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.531078 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e"} err="failed to get container status \"0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e\": rpc error: code = NotFound desc = could not find container \"0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e\": container with ID starting with 0ab926957bbbcb540033ee88df88ae3b51d9eab6de9e1f82458b1e9b7c29673e not found: ID does not exist" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.531102 4817 scope.go:117] "RemoveContainer" containerID="290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4" Nov 28 14:50:51 crc kubenswrapper[4817]: E1128 14:50:51.531407 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4\": container with ID starting with 290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4 not found: ID does not exist" containerID="290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.531433 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4"} err="failed to get container status \"290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4\": rpc error: code = NotFound desc = could not find container \"290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4\": container with ID starting with 290db3e0756554d290c41be4ec68fc389c4d43ce4688dc27bd64b7ada95c67e4 not found: ID does not exist" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652483 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a6a65e3-df90-4500-8deb-1dc661954e49-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652536 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652566 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtxt9\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-kube-api-access-rtxt9\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652586 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652607 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652648 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652670 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652687 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652703 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-config-data\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652749 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a6a65e3-df90-4500-8deb-1dc661954e49-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.652797 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.758964 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.762955 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763015 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763058 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-config-data\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763167 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a6a65e3-df90-4500-8deb-1dc661954e49-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763392 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763450 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a6a65e3-df90-4500-8deb-1dc661954e49-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763529 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763613 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtxt9\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-kube-api-access-rtxt9\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763683 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.763753 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.764078 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63b71752-f06a-4d1a-9d80-85142bbc3610" path="/var/lib/kubelet/pods/63b71752-f06a-4d1a-9d80-85142bbc3610/volumes" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.760042 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.765193 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.769392 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.769831 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-config-data\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.770155 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.772148 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a6a65e3-df90-4500-8deb-1dc661954e49-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.772677 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.774055 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a6a65e3-df90-4500-8deb-1dc661954e49-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.774844 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.775097 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a6a65e3-df90-4500-8deb-1dc661954e49-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.778047 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.787793 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtxt9\" (UniqueName: \"kubernetes.io/projected/8a6a65e3-df90-4500-8deb-1dc661954e49-kube-api-access-rtxt9\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.826126 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a6a65e3-df90-4500-8deb-1dc661954e49\") " pod="openstack/rabbitmq-server-0" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970229 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-erlang-cookie\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970385 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2d84\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-kube-api-access-f2d84\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970431 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-confd\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970465 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d340d5c2-67ed-48da-826f-e1e4c244832b-pod-info\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970500 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-plugins\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970532 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-config-data\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970646 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-server-conf\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970699 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-plugins-conf\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970772 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d340d5c2-67ed-48da-826f-e1e4c244832b-erlang-cookie-secret\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970797 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-tls\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970831 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d340d5c2-67ed-48da-826f-e1e4c244832b\" (UID: \"d340d5c2-67ed-48da-826f-e1e4c244832b\") " Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.970813 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.971219 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.972002 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.972024 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.973333 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.976819 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.977900 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d340d5c2-67ed-48da-826f-e1e4c244832b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.983342 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d340d5c2-67ed-48da-826f-e1e4c244832b-pod-info" (OuterVolumeSpecName: "pod-info") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.984868 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-kube-api-access-f2d84" (OuterVolumeSpecName: "kube-api-access-f2d84") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "kube-api-access-f2d84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:51 crc kubenswrapper[4817]: I1128 14:50:51.985932 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.008575 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-config-data" (OuterVolumeSpecName: "config-data") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.038441 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-server-conf" (OuterVolumeSpecName: "server-conf") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073632 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2d84\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-kube-api-access-f2d84\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073679 4817 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d340d5c2-67ed-48da-826f-e1e4c244832b-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073694 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073706 4817 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073750 4817 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d340d5c2-67ed-48da-826f-e1e4c244832b-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073762 4817 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d340d5c2-67ed-48da-826f-e1e4c244832b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073774 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.073808 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.089421 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d340d5c2-67ed-48da-826f-e1e4c244832b" (UID: "d340d5c2-67ed-48da-826f-e1e4c244832b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.109960 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.121995 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.175159 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.175188 4817 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d340d5c2-67ed-48da-826f-e1e4c244832b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.414286 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d340d5c2-67ed-48da-826f-e1e4c244832b","Type":"ContainerDied","Data":"a4c4f9342056520de6653dffa97f90343f2042f50c507143a12861b13ea6b4bf"} Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.414670 4817 scope.go:117] "RemoveContainer" containerID="ed4c3b1acfccbff43f85441f72bfebd43f8cba89beaf216be01366e988d4c343" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.414844 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.433212 4817 scope.go:117] "RemoveContainer" containerID="8ccf30fb29efa6e931504451a614334aecdfe48ab9668d3b04ee44037bd78dc6" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.455786 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.468231 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.484210 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:50:52 crc kubenswrapper[4817]: E1128 14:50:52.484670 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="setup-container" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.484691 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="setup-container" Nov 28 14:50:52 crc kubenswrapper[4817]: E1128 14:50:52.484707 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="rabbitmq" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.484714 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="rabbitmq" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.484957 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" containerName="rabbitmq" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.486157 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.490576 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.490591 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.490671 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.490732 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qh99g" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.490591 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.490831 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.491214 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.503015 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.586885 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.586938 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.586972 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587064 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587085 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587154 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587175 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5cbe25b-ed99-4853-a3bd-726f25343f6a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587235 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587254 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587283 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzlvt\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-kube-api-access-pzlvt\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.587299 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5cbe25b-ed99-4853-a3bd-726f25343f6a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.618645 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.688779 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzlvt\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-kube-api-access-pzlvt\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689126 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5cbe25b-ed99-4853-a3bd-726f25343f6a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689180 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689229 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689279 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689364 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689404 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.689987 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690034 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5cbe25b-ed99-4853-a3bd-726f25343f6a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690108 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690183 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690453 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690631 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690651 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.690762 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.691900 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.692577 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5cbe25b-ed99-4853-a3bd-726f25343f6a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.694417 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.695125 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5cbe25b-ed99-4853-a3bd-726f25343f6a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.695245 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5cbe25b-ed99-4853-a3bd-726f25343f6a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.699804 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.711351 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzlvt\" (UniqueName: \"kubernetes.io/projected/a5cbe25b-ed99-4853-a3bd-726f25343f6a-kube-api-access-pzlvt\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.727128 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5cbe25b-ed99-4853-a3bd-726f25343f6a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:52 crc kubenswrapper[4817]: I1128 14:50:52.816439 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:50:53 crc kubenswrapper[4817]: I1128 14:50:53.262541 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 14:50:53 crc kubenswrapper[4817]: I1128 14:50:53.432045 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5cbe25b-ed99-4853-a3bd-726f25343f6a","Type":"ContainerStarted","Data":"55e9a4ef9afdfae8800229be6b7f6cd9a9f4b4ef5fa94eb435dc328efdc50f5f"} Nov 28 14:50:53 crc kubenswrapper[4817]: I1128 14:50:53.434651 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a6a65e3-df90-4500-8deb-1dc661954e49","Type":"ContainerStarted","Data":"f1def7c4df5a7eabcc2382fbb3d57e417455b46e77a484c956de3b0597e82c64"} Nov 28 14:50:53 crc kubenswrapper[4817]: I1128 14:50:53.748660 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d340d5c2-67ed-48da-826f-e1e4c244832b" path="/var/lib/kubelet/pods/d340d5c2-67ed-48da-826f-e1e4c244832b/volumes" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.280487 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-g4pc8"] Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.282091 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.284711 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.308595 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-g4pc8"] Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.428373 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.428463 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.428501 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-config\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.428553 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.428812 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.428877 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb48h\" (UniqueName: \"kubernetes.io/projected/cc1c4440-f241-4159-b5fb-a772622d5065-kube-api-access-qb48h\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.429008 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.452228 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a6a65e3-df90-4500-8deb-1dc661954e49","Type":"ContainerStarted","Data":"1d7ec20006a2d59887ac958a34dd49a9c8cd498ac4fe56cae95953de5990e24d"} Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.530997 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.531111 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-config\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.531254 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.531318 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.531354 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb48h\" (UniqueName: \"kubernetes.io/projected/cc1c4440-f241-4159-b5fb-a772622d5065-kube-api-access-qb48h\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.531401 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.531512 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.532517 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.532975 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-config\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.533318 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.533762 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.534081 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.535861 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.566218 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb48h\" (UniqueName: \"kubernetes.io/projected/cc1c4440-f241-4159-b5fb-a772622d5065-kube-api-access-qb48h\") pod \"dnsmasq-dns-7d84b4d45c-g4pc8\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:54 crc kubenswrapper[4817]: I1128 14:50:54.618442 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:56 crc kubenswrapper[4817]: I1128 14:50:55.131207 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-g4pc8"] Nov 28 14:50:56 crc kubenswrapper[4817]: I1128 14:50:55.466605 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" event={"ID":"cc1c4440-f241-4159-b5fb-a772622d5065","Type":"ContainerStarted","Data":"ea37749165114b217956c5f7f3003c8b6a6b19cda342c9af46597bc487964a9b"} Nov 28 14:50:56 crc kubenswrapper[4817]: I1128 14:50:55.469401 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5cbe25b-ed99-4853-a3bd-726f25343f6a","Type":"ContainerStarted","Data":"2c79f91a99e91a8ae6177a3b1452ac898dbe230cd55756c99a896f318de78249"} Nov 28 14:50:56 crc kubenswrapper[4817]: I1128 14:50:56.479136 4817 generic.go:334] "Generic (PLEG): container finished" podID="cc1c4440-f241-4159-b5fb-a772622d5065" containerID="c2ac7fd5231dda57476a38e3d2a8353c4984fbb75ac597f2b2db67f43e4381cf" exitCode=0 Nov 28 14:50:56 crc kubenswrapper[4817]: I1128 14:50:56.479825 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" event={"ID":"cc1c4440-f241-4159-b5fb-a772622d5065","Type":"ContainerDied","Data":"c2ac7fd5231dda57476a38e3d2a8353c4984fbb75ac597f2b2db67f43e4381cf"} Nov 28 14:50:57 crc kubenswrapper[4817]: I1128 14:50:57.490442 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" event={"ID":"cc1c4440-f241-4159-b5fb-a772622d5065","Type":"ContainerStarted","Data":"949e2affeb68873e59db910e9ea2585a115dc535b5390fe30a2cdbe816110bed"} Nov 28 14:50:57 crc kubenswrapper[4817]: I1128 14:50:57.490821 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:50:57 crc kubenswrapper[4817]: I1128 14:50:57.527769 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" podStartSLOduration=3.527708619 podStartE2EDuration="3.527708619s" podCreationTimestamp="2025-11-28 14:50:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:50:57.509599709 +0000 UTC m=+1280.097577965" watchObservedRunningTime="2025-11-28 14:50:57.527708619 +0000 UTC m=+1280.115686915" Nov 28 14:51:04 crc kubenswrapper[4817]: I1128 14:51:04.621138 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:51:04 crc kubenswrapper[4817]: I1128 14:51:04.693220 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-f95c7"] Nov 28 14:51:04 crc kubenswrapper[4817]: I1128 14:51:04.693470 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerName="dnsmasq-dns" containerID="cri-o://5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778" gracePeriod=10 Nov 28 14:51:04 crc kubenswrapper[4817]: I1128 14:51:04.911089 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-f8k7g"] Nov 28 14:51:04 crc kubenswrapper[4817]: I1128 14:51:04.913205 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:04 crc kubenswrapper[4817]: I1128 14:51:04.929850 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-f8k7g"] Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054160 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054208 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054238 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054375 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd2st\" (UniqueName: \"kubernetes.io/projected/6ecb4997-6b98-4f7b-ad43-e02295ffe723-kube-api-access-bd2st\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054477 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054615 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.054705 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-config\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.157170 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-config\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.157530 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.157628 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.157706 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.157871 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd2st\" (UniqueName: \"kubernetes.io/projected/6ecb4997-6b98-4f7b-ad43-e02295ffe723-kube-api-access-bd2st\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.157953 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.158050 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.158743 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.158804 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.159028 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.159492 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.159669 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.159873 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ecb4997-6b98-4f7b-ad43-e02295ffe723-config\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.191752 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd2st\" (UniqueName: \"kubernetes.io/projected/6ecb4997-6b98-4f7b-ad43-e02295ffe723-kube-api-access-bd2st\") pod \"dnsmasq-dns-6f6df4f56c-f8k7g\" (UID: \"6ecb4997-6b98-4f7b-ad43-e02295ffe723\") " pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.232809 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.358504 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.464506 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stcnk\" (UniqueName: \"kubernetes.io/projected/a1097a83-dca3-4ad8-9793-8a68affb77da-kube-api-access-stcnk\") pod \"a1097a83-dca3-4ad8-9793-8a68affb77da\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.465347 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-swift-storage-0\") pod \"a1097a83-dca3-4ad8-9793-8a68affb77da\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.465443 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-svc\") pod \"a1097a83-dca3-4ad8-9793-8a68affb77da\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.465528 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-config\") pod \"a1097a83-dca3-4ad8-9793-8a68affb77da\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.465961 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-nb\") pod \"a1097a83-dca3-4ad8-9793-8a68affb77da\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.466051 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-sb\") pod \"a1097a83-dca3-4ad8-9793-8a68affb77da\" (UID: \"a1097a83-dca3-4ad8-9793-8a68affb77da\") " Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.469804 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1097a83-dca3-4ad8-9793-8a68affb77da-kube-api-access-stcnk" (OuterVolumeSpecName: "kube-api-access-stcnk") pod "a1097a83-dca3-4ad8-9793-8a68affb77da" (UID: "a1097a83-dca3-4ad8-9793-8a68affb77da"). InnerVolumeSpecName "kube-api-access-stcnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.537360 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-config" (OuterVolumeSpecName: "config") pod "a1097a83-dca3-4ad8-9793-8a68affb77da" (UID: "a1097a83-dca3-4ad8-9793-8a68affb77da"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.541355 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a1097a83-dca3-4ad8-9793-8a68affb77da" (UID: "a1097a83-dca3-4ad8-9793-8a68affb77da"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.542016 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a1097a83-dca3-4ad8-9793-8a68affb77da" (UID: "a1097a83-dca3-4ad8-9793-8a68affb77da"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.554764 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1097a83-dca3-4ad8-9793-8a68affb77da" (UID: "a1097a83-dca3-4ad8-9793-8a68affb77da"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.555452 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a1097a83-dca3-4ad8-9793-8a68affb77da" (UID: "a1097a83-dca3-4ad8-9793-8a68affb77da"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.568994 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stcnk\" (UniqueName: \"kubernetes.io/projected/a1097a83-dca3-4ad8-9793-8a68affb77da-kube-api-access-stcnk\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.569024 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.569038 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.569050 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.569062 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.569074 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1097a83-dca3-4ad8-9793-8a68affb77da-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.572591 4817 generic.go:334] "Generic (PLEG): container finished" podID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerID="5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778" exitCode=0 Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.572631 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" event={"ID":"a1097a83-dca3-4ad8-9793-8a68affb77da","Type":"ContainerDied","Data":"5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778"} Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.572654 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.572668 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-f95c7" event={"ID":"a1097a83-dca3-4ad8-9793-8a68affb77da","Type":"ContainerDied","Data":"8e1c6a71cc575ab8a014d301a176c3a50f2e3c0f2c3b938f8d4fc2c2a732d32e"} Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.572689 4817 scope.go:117] "RemoveContainer" containerID="5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.605958 4817 scope.go:117] "RemoveContainer" containerID="5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.608041 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-f95c7"] Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.622084 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-f95c7"] Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.651788 4817 scope.go:117] "RemoveContainer" containerID="5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778" Nov 28 14:51:05 crc kubenswrapper[4817]: E1128 14:51:05.653921 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778\": container with ID starting with 5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778 not found: ID does not exist" containerID="5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.653993 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778"} err="failed to get container status \"5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778\": rpc error: code = NotFound desc = could not find container \"5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778\": container with ID starting with 5baf71bf571b406bdbf4780a136bf5df7135bbf256aa15cc10f14f7431bdf778 not found: ID does not exist" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.654032 4817 scope.go:117] "RemoveContainer" containerID="5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e" Nov 28 14:51:05 crc kubenswrapper[4817]: E1128 14:51:05.655013 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e\": container with ID starting with 5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e not found: ID does not exist" containerID="5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.655087 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e"} err="failed to get container status \"5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e\": rpc error: code = NotFound desc = could not find container \"5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e\": container with ID starting with 5c9ecdd2a0454a042668097bd7c0f3a7e3d401f097e2b02f6e804910d5f2dc3e not found: ID does not exist" Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.710377 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-f8k7g"] Nov 28 14:51:05 crc kubenswrapper[4817]: W1128 14:51:05.719002 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ecb4997_6b98_4f7b_ad43_e02295ffe723.slice/crio-44164e6bbdc659e1f83171b4088a35fe42a12f518d96841bcd010a770ee1d3a9 WatchSource:0}: Error finding container 44164e6bbdc659e1f83171b4088a35fe42a12f518d96841bcd010a770ee1d3a9: Status 404 returned error can't find the container with id 44164e6bbdc659e1f83171b4088a35fe42a12f518d96841bcd010a770ee1d3a9 Nov 28 14:51:05 crc kubenswrapper[4817]: I1128 14:51:05.750515 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" path="/var/lib/kubelet/pods/a1097a83-dca3-4ad8-9793-8a68affb77da/volumes" Nov 28 14:51:06 crc kubenswrapper[4817]: I1128 14:51:06.585018 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" event={"ID":"6ecb4997-6b98-4f7b-ad43-e02295ffe723","Type":"ContainerStarted","Data":"a18689c979b1602da71817b99d065a41f32d3318132061827d8a597f083940b3"} Nov 28 14:51:06 crc kubenswrapper[4817]: I1128 14:51:06.585773 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" event={"ID":"6ecb4997-6b98-4f7b-ad43-e02295ffe723","Type":"ContainerStarted","Data":"44164e6bbdc659e1f83171b4088a35fe42a12f518d96841bcd010a770ee1d3a9"} Nov 28 14:51:07 crc kubenswrapper[4817]: I1128 14:51:07.594570 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ecb4997-6b98-4f7b-ad43-e02295ffe723" containerID="a18689c979b1602da71817b99d065a41f32d3318132061827d8a597f083940b3" exitCode=0 Nov 28 14:51:07 crc kubenswrapper[4817]: I1128 14:51:07.594650 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" event={"ID":"6ecb4997-6b98-4f7b-ad43-e02295ffe723","Type":"ContainerDied","Data":"a18689c979b1602da71817b99d065a41f32d3318132061827d8a597f083940b3"} Nov 28 14:51:08 crc kubenswrapper[4817]: I1128 14:51:08.609132 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" event={"ID":"6ecb4997-6b98-4f7b-ad43-e02295ffe723","Type":"ContainerStarted","Data":"8890b58b4594456d93bd04822b729dc899c8d6c3ce243fa9a13a5b2309caa660"} Nov 28 14:51:08 crc kubenswrapper[4817]: I1128 14:51:08.609806 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:08 crc kubenswrapper[4817]: I1128 14:51:08.643275 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" podStartSLOduration=4.643260167 podStartE2EDuration="4.643260167s" podCreationTimestamp="2025-11-28 14:51:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:51:08.636980871 +0000 UTC m=+1291.224959197" watchObservedRunningTime="2025-11-28 14:51:08.643260167 +0000 UTC m=+1291.231238423" Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.234863 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-f8k7g" Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.344874 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-g4pc8"] Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.345112 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" containerName="dnsmasq-dns" containerID="cri-o://949e2affeb68873e59db910e9ea2585a115dc535b5390fe30a2cdbe816110bed" gracePeriod=10 Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.693625 4817 generic.go:334] "Generic (PLEG): container finished" podID="cc1c4440-f241-4159-b5fb-a772622d5065" containerID="949e2affeb68873e59db910e9ea2585a115dc535b5390fe30a2cdbe816110bed" exitCode=0 Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.693764 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" event={"ID":"cc1c4440-f241-4159-b5fb-a772622d5065","Type":"ContainerDied","Data":"949e2affeb68873e59db910e9ea2585a115dc535b5390fe30a2cdbe816110bed"} Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.832917 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976340 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-swift-storage-0\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976482 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-openstack-edpm-ipam\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976510 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-sb\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976571 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-nb\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976653 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-svc\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976760 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-config\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.976842 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb48h\" (UniqueName: \"kubernetes.io/projected/cc1c4440-f241-4159-b5fb-a772622d5065-kube-api-access-qb48h\") pod \"cc1c4440-f241-4159-b5fb-a772622d5065\" (UID: \"cc1c4440-f241-4159-b5fb-a772622d5065\") " Nov 28 14:51:15 crc kubenswrapper[4817]: I1128 14:51:15.986113 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc1c4440-f241-4159-b5fb-a772622d5065-kube-api-access-qb48h" (OuterVolumeSpecName: "kube-api-access-qb48h") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "kube-api-access-qb48h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.036279 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.038200 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.039182 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.056399 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.057599 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.066047 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-config" (OuterVolumeSpecName: "config") pod "cc1c4440-f241-4159-b5fb-a772622d5065" (UID: "cc1c4440-f241-4159-b5fb-a772622d5065"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080161 4817 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080192 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080202 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb48h\" (UniqueName: \"kubernetes.io/projected/cc1c4440-f241-4159-b5fb-a772622d5065-kube-api-access-qb48h\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080213 4817 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080222 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080230 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.080239 4817 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc1c4440-f241-4159-b5fb-a772622d5065-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.704624 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" event={"ID":"cc1c4440-f241-4159-b5fb-a772622d5065","Type":"ContainerDied","Data":"ea37749165114b217956c5f7f3003c8b6a6b19cda342c9af46597bc487964a9b"} Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.704668 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-g4pc8" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.704669 4817 scope.go:117] "RemoveContainer" containerID="949e2affeb68873e59db910e9ea2585a115dc535b5390fe30a2cdbe816110bed" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.739712 4817 scope.go:117] "RemoveContainer" containerID="c2ac7fd5231dda57476a38e3d2a8353c4984fbb75ac597f2b2db67f43e4381cf" Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.751026 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-g4pc8"] Nov 28 14:51:16 crc kubenswrapper[4817]: I1128 14:51:16.764756 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-g4pc8"] Nov 28 14:51:17 crc kubenswrapper[4817]: I1128 14:51:17.762487 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" path="/var/lib/kubelet/pods/cc1c4440-f241-4159-b5fb-a772622d5065/volumes" Nov 28 14:51:26 crc kubenswrapper[4817]: I1128 14:51:26.794377 4817 generic.go:334] "Generic (PLEG): container finished" podID="8a6a65e3-df90-4500-8deb-1dc661954e49" containerID="1d7ec20006a2d59887ac958a34dd49a9c8cd498ac4fe56cae95953de5990e24d" exitCode=0 Nov 28 14:51:26 crc kubenswrapper[4817]: I1128 14:51:26.794506 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a6a65e3-df90-4500-8deb-1dc661954e49","Type":"ContainerDied","Data":"1d7ec20006a2d59887ac958a34dd49a9c8cd498ac4fe56cae95953de5990e24d"} Nov 28 14:51:27 crc kubenswrapper[4817]: I1128 14:51:27.807260 4817 generic.go:334] "Generic (PLEG): container finished" podID="a5cbe25b-ed99-4853-a3bd-726f25343f6a" containerID="2c79f91a99e91a8ae6177a3b1452ac898dbe230cd55756c99a896f318de78249" exitCode=0 Nov 28 14:51:27 crc kubenswrapper[4817]: I1128 14:51:27.807348 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5cbe25b-ed99-4853-a3bd-726f25343f6a","Type":"ContainerDied","Data":"2c79f91a99e91a8ae6177a3b1452ac898dbe230cd55756c99a896f318de78249"} Nov 28 14:51:27 crc kubenswrapper[4817]: I1128 14:51:27.810241 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a6a65e3-df90-4500-8deb-1dc661954e49","Type":"ContainerStarted","Data":"0280257c987a61bfae6c06d99dbdb56b8cb08217c6aa2a3d19335eb5168e75c8"} Nov 28 14:51:27 crc kubenswrapper[4817]: I1128 14:51:27.810442 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 28 14:51:27 crc kubenswrapper[4817]: I1128 14:51:27.880461 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.880436843 podStartE2EDuration="36.880436843s" podCreationTimestamp="2025-11-28 14:50:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:51:27.868321151 +0000 UTC m=+1310.456299447" watchObservedRunningTime="2025-11-28 14:51:27.880436843 +0000 UTC m=+1310.468415119" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.423860 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd"] Nov 28 14:51:28 crc kubenswrapper[4817]: E1128 14:51:28.424314 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerName="dnsmasq-dns" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.424339 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerName="dnsmasq-dns" Nov 28 14:51:28 crc kubenswrapper[4817]: E1128 14:51:28.424369 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" containerName="init" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.424400 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" containerName="init" Nov 28 14:51:28 crc kubenswrapper[4817]: E1128 14:51:28.424439 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" containerName="dnsmasq-dns" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.424448 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" containerName="dnsmasq-dns" Nov 28 14:51:28 crc kubenswrapper[4817]: E1128 14:51:28.424463 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerName="init" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.424471 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerName="init" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.424828 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1c4440-f241-4159-b5fb-a772622d5065" containerName="dnsmasq-dns" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.424991 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1097a83-dca3-4ad8-9793-8a68affb77da" containerName="dnsmasq-dns" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.425783 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.427780 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.427954 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.428909 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.429231 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.446747 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd"] Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.529928 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.530131 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.530236 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4sq\" (UniqueName: \"kubernetes.io/projected/02fe616c-2dd3-412b-a8e1-9af3150fc92c-kube-api-access-wq4sq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.530576 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.632123 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.632238 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.632285 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4sq\" (UniqueName: \"kubernetes.io/projected/02fe616c-2dd3-412b-a8e1-9af3150fc92c-kube-api-access-wq4sq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.632323 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.637500 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.637650 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.641815 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.654608 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4sq\" (UniqueName: \"kubernetes.io/projected/02fe616c-2dd3-412b-a8e1-9af3150fc92c-kube-api-access-wq4sq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.742492 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.836949 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5cbe25b-ed99-4853-a3bd-726f25343f6a","Type":"ContainerStarted","Data":"52a95b972da5806b1e81ef80bd3ffe581c265548f332ac79a430dd92867600ba"} Nov 28 14:51:28 crc kubenswrapper[4817]: I1128 14:51:28.838423 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:51:29 crc kubenswrapper[4817]: I1128 14:51:29.317905 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.31787912 podStartE2EDuration="37.31787912s" podCreationTimestamp="2025-11-28 14:50:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:51:28.864359594 +0000 UTC m=+1311.452337860" watchObservedRunningTime="2025-11-28 14:51:29.31787912 +0000 UTC m=+1311.905857386" Nov 28 14:51:29 crc kubenswrapper[4817]: I1128 14:51:29.323289 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd"] Nov 28 14:51:29 crc kubenswrapper[4817]: W1128 14:51:29.345895 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02fe616c_2dd3_412b_a8e1_9af3150fc92c.slice/crio-0f09bc5429c02b5a478a260f6bbe960aac804496780012305e6d4ec914912292 WatchSource:0}: Error finding container 0f09bc5429c02b5a478a260f6bbe960aac804496780012305e6d4ec914912292: Status 404 returned error can't find the container with id 0f09bc5429c02b5a478a260f6bbe960aac804496780012305e6d4ec914912292 Nov 28 14:51:29 crc kubenswrapper[4817]: I1128 14:51:29.849100 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" event={"ID":"02fe616c-2dd3-412b-a8e1-9af3150fc92c","Type":"ContainerStarted","Data":"0f09bc5429c02b5a478a260f6bbe960aac804496780012305e6d4ec914912292"} Nov 28 14:51:39 crc kubenswrapper[4817]: I1128 14:51:39.834574 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:51:40 crc kubenswrapper[4817]: I1128 14:51:40.976506 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" event={"ID":"02fe616c-2dd3-412b-a8e1-9af3150fc92c","Type":"ContainerStarted","Data":"8e27bdd175754e2e9af467c2446b5423674ced50b59b3354aa8eeadda33bbcdb"} Nov 28 14:51:40 crc kubenswrapper[4817]: I1128 14:51:40.993436 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" podStartSLOduration=2.5085852060000002 podStartE2EDuration="12.993415116s" podCreationTimestamp="2025-11-28 14:51:28 +0000 UTC" firstStartedPulling="2025-11-28 14:51:29.347417107 +0000 UTC m=+1311.935395373" lastFinishedPulling="2025-11-28 14:51:39.832247017 +0000 UTC m=+1322.420225283" observedRunningTime="2025-11-28 14:51:40.990012181 +0000 UTC m=+1323.577990467" watchObservedRunningTime="2025-11-28 14:51:40.993415116 +0000 UTC m=+1323.581393392" Nov 28 14:51:42 crc kubenswrapper[4817]: I1128 14:51:42.126030 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 28 14:51:42 crc kubenswrapper[4817]: I1128 14:51:42.819942 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 28 14:51:52 crc kubenswrapper[4817]: I1128 14:51:52.095210 4817 generic.go:334] "Generic (PLEG): container finished" podID="02fe616c-2dd3-412b-a8e1-9af3150fc92c" containerID="8e27bdd175754e2e9af467c2446b5423674ced50b59b3354aa8eeadda33bbcdb" exitCode=0 Nov 28 14:51:52 crc kubenswrapper[4817]: I1128 14:51:52.095338 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" event={"ID":"02fe616c-2dd3-412b-a8e1-9af3150fc92c","Type":"ContainerDied","Data":"8e27bdd175754e2e9af467c2446b5423674ced50b59b3354aa8eeadda33bbcdb"} Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.519981 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.614007 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq4sq\" (UniqueName: \"kubernetes.io/projected/02fe616c-2dd3-412b-a8e1-9af3150fc92c-kube-api-access-wq4sq\") pod \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.614137 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-repo-setup-combined-ca-bundle\") pod \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.614289 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-inventory\") pod \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.614401 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-ssh-key\") pod \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\" (UID: \"02fe616c-2dd3-412b-a8e1-9af3150fc92c\") " Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.621994 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fe616c-2dd3-412b-a8e1-9af3150fc92c-kube-api-access-wq4sq" (OuterVolumeSpecName: "kube-api-access-wq4sq") pod "02fe616c-2dd3-412b-a8e1-9af3150fc92c" (UID: "02fe616c-2dd3-412b-a8e1-9af3150fc92c"). InnerVolumeSpecName "kube-api-access-wq4sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.622006 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "02fe616c-2dd3-412b-a8e1-9af3150fc92c" (UID: "02fe616c-2dd3-412b-a8e1-9af3150fc92c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.640065 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "02fe616c-2dd3-412b-a8e1-9af3150fc92c" (UID: "02fe616c-2dd3-412b-a8e1-9af3150fc92c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.643906 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-inventory" (OuterVolumeSpecName: "inventory") pod "02fe616c-2dd3-412b-a8e1-9af3150fc92c" (UID: "02fe616c-2dd3-412b-a8e1-9af3150fc92c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.717252 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.717285 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.717296 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq4sq\" (UniqueName: \"kubernetes.io/projected/02fe616c-2dd3-412b-a8e1-9af3150fc92c-kube-api-access-wq4sq\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:53 crc kubenswrapper[4817]: I1128 14:51:53.717307 4817 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02fe616c-2dd3-412b-a8e1-9af3150fc92c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.127818 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" event={"ID":"02fe616c-2dd3-412b-a8e1-9af3150fc92c","Type":"ContainerDied","Data":"0f09bc5429c02b5a478a260f6bbe960aac804496780012305e6d4ec914912292"} Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.127866 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f09bc5429c02b5a478a260f6bbe960aac804496780012305e6d4ec914912292" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.127934 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.269028 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk"] Nov 28 14:51:54 crc kubenswrapper[4817]: E1128 14:51:54.269433 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fe616c-2dd3-412b-a8e1-9af3150fc92c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.269458 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fe616c-2dd3-412b-a8e1-9af3150fc92c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.283176 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fe616c-2dd3-412b-a8e1-9af3150fc92c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.286533 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk"] Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.286638 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.290207 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.290536 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.293346 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.293396 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.430548 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.430614 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr6gm\" (UniqueName: \"kubernetes.io/projected/14f43d11-6db6-461c-9279-22fdd3e286b8-kube-api-access-fr6gm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.430748 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.532809 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.532884 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr6gm\" (UniqueName: \"kubernetes.io/projected/14f43d11-6db6-461c-9279-22fdd3e286b8-kube-api-access-fr6gm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.532933 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.537011 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.552846 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.554126 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr6gm\" (UniqueName: \"kubernetes.io/projected/14f43d11-6db6-461c-9279-22fdd3e286b8-kube-api-access-fr6gm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zzxtk\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:54 crc kubenswrapper[4817]: I1128 14:51:54.611578 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:51:55 crc kubenswrapper[4817]: I1128 14:51:55.223303 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk"] Nov 28 14:51:56 crc kubenswrapper[4817]: I1128 14:51:56.149949 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" event={"ID":"14f43d11-6db6-461c-9279-22fdd3e286b8","Type":"ContainerStarted","Data":"6f3ffef9f7f9acbd58d6e2fff95fd041e3642ea7280d8ad2b9ac2382e3d717e7"} Nov 28 14:51:56 crc kubenswrapper[4817]: I1128 14:51:56.150008 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" event={"ID":"14f43d11-6db6-461c-9279-22fdd3e286b8","Type":"ContainerStarted","Data":"22b939a1e4433b41897ccb6149cd8f8c027dab5a0d3ee179649464938a23ff4c"} Nov 28 14:51:56 crc kubenswrapper[4817]: I1128 14:51:56.166559 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" podStartSLOduration=1.736514481 podStartE2EDuration="2.166542552s" podCreationTimestamp="2025-11-28 14:51:54 +0000 UTC" firstStartedPulling="2025-11-28 14:51:55.228807313 +0000 UTC m=+1337.816785599" lastFinishedPulling="2025-11-28 14:51:55.658835394 +0000 UTC m=+1338.246813670" observedRunningTime="2025-11-28 14:51:56.165743522 +0000 UTC m=+1338.753721798" watchObservedRunningTime="2025-11-28 14:51:56.166542552 +0000 UTC m=+1338.754520818" Nov 28 14:51:59 crc kubenswrapper[4817]: I1128 14:51:59.189555 4817 generic.go:334] "Generic (PLEG): container finished" podID="14f43d11-6db6-461c-9279-22fdd3e286b8" containerID="6f3ffef9f7f9acbd58d6e2fff95fd041e3642ea7280d8ad2b9ac2382e3d717e7" exitCode=0 Nov 28 14:51:59 crc kubenswrapper[4817]: I1128 14:51:59.189660 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" event={"ID":"14f43d11-6db6-461c-9279-22fdd3e286b8","Type":"ContainerDied","Data":"6f3ffef9f7f9acbd58d6e2fff95fd041e3642ea7280d8ad2b9ac2382e3d717e7"} Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.603412 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.645665 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-inventory\") pod \"14f43d11-6db6-461c-9279-22fdd3e286b8\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.645810 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-ssh-key\") pod \"14f43d11-6db6-461c-9279-22fdd3e286b8\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.645902 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr6gm\" (UniqueName: \"kubernetes.io/projected/14f43d11-6db6-461c-9279-22fdd3e286b8-kube-api-access-fr6gm\") pod \"14f43d11-6db6-461c-9279-22fdd3e286b8\" (UID: \"14f43d11-6db6-461c-9279-22fdd3e286b8\") " Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.673336 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f43d11-6db6-461c-9279-22fdd3e286b8-kube-api-access-fr6gm" (OuterVolumeSpecName: "kube-api-access-fr6gm") pod "14f43d11-6db6-461c-9279-22fdd3e286b8" (UID: "14f43d11-6db6-461c-9279-22fdd3e286b8"). InnerVolumeSpecName "kube-api-access-fr6gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.698169 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-inventory" (OuterVolumeSpecName: "inventory") pod "14f43d11-6db6-461c-9279-22fdd3e286b8" (UID: "14f43d11-6db6-461c-9279-22fdd3e286b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.705895 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "14f43d11-6db6-461c-9279-22fdd3e286b8" (UID: "14f43d11-6db6-461c-9279-22fdd3e286b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.748386 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.748414 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr6gm\" (UniqueName: \"kubernetes.io/projected/14f43d11-6db6-461c-9279-22fdd3e286b8-kube-api-access-fr6gm\") on node \"crc\" DevicePath \"\"" Nov 28 14:52:00 crc kubenswrapper[4817]: I1128 14:52:00.748424 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f43d11-6db6-461c-9279-22fdd3e286b8-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.218157 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" event={"ID":"14f43d11-6db6-461c-9279-22fdd3e286b8","Type":"ContainerDied","Data":"22b939a1e4433b41897ccb6149cd8f8c027dab5a0d3ee179649464938a23ff4c"} Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.218207 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zzxtk" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.218194 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22b939a1e4433b41897ccb6149cd8f8c027dab5a0d3ee179649464938a23ff4c" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.312591 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q"] Nov 28 14:52:01 crc kubenswrapper[4817]: E1128 14:52:01.313130 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f43d11-6db6-461c-9279-22fdd3e286b8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.313165 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f43d11-6db6-461c-9279-22fdd3e286b8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.313414 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f43d11-6db6-461c-9279-22fdd3e286b8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.314456 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.316320 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.316357 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.316595 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.318171 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.322392 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q"] Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.362223 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz272\" (UniqueName: \"kubernetes.io/projected/3200157e-779d-485c-803e-1b0d092c02bb-kube-api-access-zz272\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.362281 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.362576 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.362728 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.464379 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.464459 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.464520 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz272\" (UniqueName: \"kubernetes.io/projected/3200157e-779d-485c-803e-1b0d092c02bb-kube-api-access-zz272\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.464554 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.469362 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.469628 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.469992 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.482441 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz272\" (UniqueName: \"kubernetes.io/projected/3200157e-779d-485c-803e-1b0d092c02bb-kube-api-access-zz272\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:01 crc kubenswrapper[4817]: I1128 14:52:01.635700 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:52:02 crc kubenswrapper[4817]: I1128 14:52:02.242549 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q"] Nov 28 14:52:03 crc kubenswrapper[4817]: I1128 14:52:03.238774 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" event={"ID":"3200157e-779d-485c-803e-1b0d092c02bb","Type":"ContainerStarted","Data":"4019c41d50178edb39d8d3c852142613128cf181ad8c1dc3ce214dbcdba062ce"} Nov 28 14:52:03 crc kubenswrapper[4817]: I1128 14:52:03.239119 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" event={"ID":"3200157e-779d-485c-803e-1b0d092c02bb","Type":"ContainerStarted","Data":"d24ced0ca42b845f98727767ef44dd4ab3677d206b9e088f7d581d9e73918369"} Nov 28 14:52:03 crc kubenswrapper[4817]: I1128 14:52:03.262939 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" podStartSLOduration=1.8120966539999999 podStartE2EDuration="2.262921694s" podCreationTimestamp="2025-11-28 14:52:01 +0000 UTC" firstStartedPulling="2025-11-28 14:52:02.231803866 +0000 UTC m=+1344.819782122" lastFinishedPulling="2025-11-28 14:52:02.682628846 +0000 UTC m=+1345.270607162" observedRunningTime="2025-11-28 14:52:03.257088708 +0000 UTC m=+1345.845067004" watchObservedRunningTime="2025-11-28 14:52:03.262921694 +0000 UTC m=+1345.850899960" Nov 28 14:52:10 crc kubenswrapper[4817]: I1128 14:52:10.631545 4817 scope.go:117] "RemoveContainer" containerID="9e8c095cafb5dfbb3c775ef8f042420e9d9c421fd42257b5f4fc354ea195c940" Nov 28 14:52:10 crc kubenswrapper[4817]: I1128 14:52:10.665995 4817 scope.go:117] "RemoveContainer" containerID="e00dd23c0aaeb25c7291bfe0d3d52cc13f85f7a4254c3df7cfd0281423e2c128" Nov 28 14:52:24 crc kubenswrapper[4817]: I1128 14:52:24.045691 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:52:24 crc kubenswrapper[4817]: I1128 14:52:24.046498 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:52:54 crc kubenswrapper[4817]: I1128 14:52:54.045081 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:52:54 crc kubenswrapper[4817]: I1128 14:52:54.046029 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:53:10 crc kubenswrapper[4817]: I1128 14:53:10.800099 4817 scope.go:117] "RemoveContainer" containerID="a5fd050353186666463bdba9e4211db37581f9eb567c83012d2066140082892a" Nov 28 14:53:10 crc kubenswrapper[4817]: I1128 14:53:10.839470 4817 scope.go:117] "RemoveContainer" containerID="224926ac3023b14675f7a87fa20d95ea669b40299c9b456a8fdf8fac8bb58bc6" Nov 28 14:53:24 crc kubenswrapper[4817]: I1128 14:53:24.044649 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:53:24 crc kubenswrapper[4817]: I1128 14:53:24.044962 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:53:24 crc kubenswrapper[4817]: I1128 14:53:24.045003 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:53:24 crc kubenswrapper[4817]: I1128 14:53:24.045655 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:53:24 crc kubenswrapper[4817]: I1128 14:53:24.045698 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6" gracePeriod=600 Nov 28 14:53:24 crc kubenswrapper[4817]: E1128 14:53:24.117758 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f2c731e_6fa9_4e0c_81dd_115f13fa4dfc.slice/crio-d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f2c731e_6fa9_4e0c_81dd_115f13fa4dfc.slice/crio-conmon-d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6.scope\": RecentStats: unable to find data in memory cache]" Nov 28 14:53:25 crc kubenswrapper[4817]: I1128 14:53:25.110630 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6" exitCode=0 Nov 28 14:53:25 crc kubenswrapper[4817]: I1128 14:53:25.110697 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6"} Nov 28 14:53:25 crc kubenswrapper[4817]: I1128 14:53:25.111456 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290"} Nov 28 14:53:25 crc kubenswrapper[4817]: I1128 14:53:25.111493 4817 scope.go:117] "RemoveContainer" containerID="6233c40182fb7115b99ef7443b2744ab4c0eae4c9222a791b73fbfc8281a9e58" Nov 28 14:54:10 crc kubenswrapper[4817]: I1128 14:54:10.910238 4817 scope.go:117] "RemoveContainer" containerID="98b76e96e43924f6b4865883de4166204f1ee77ce977c706ea40861adb11191f" Nov 28 14:54:10 crc kubenswrapper[4817]: I1128 14:54:10.959235 4817 scope.go:117] "RemoveContainer" containerID="cacb69c5846a15b08b4240e267546dc031e7dd959725b73c99f209add92bf579" Nov 28 14:54:10 crc kubenswrapper[4817]: I1128 14:54:10.993055 4817 scope.go:117] "RemoveContainer" containerID="cc5ce7039dec187f7ab155832429a56421d01dca1b22c0f61f8d420de16056fe" Nov 28 14:54:11 crc kubenswrapper[4817]: I1128 14:54:11.022148 4817 scope.go:117] "RemoveContainer" containerID="9c043a562b41a41ac20a4fc97fc6c7620409f0425c95c6f82e3900204828bea3" Nov 28 14:54:11 crc kubenswrapper[4817]: I1128 14:54:11.055671 4817 scope.go:117] "RemoveContainer" containerID="23d817e3e29bb98b2041aefa13adf7217406c3e7c421863d97ec0f86a7b51d0c" Nov 28 14:55:24 crc kubenswrapper[4817]: I1128 14:55:24.045125 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:55:24 crc kubenswrapper[4817]: I1128 14:55:24.045730 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:55:29 crc kubenswrapper[4817]: I1128 14:55:29.167076 4817 generic.go:334] "Generic (PLEG): container finished" podID="3200157e-779d-485c-803e-1b0d092c02bb" containerID="4019c41d50178edb39d8d3c852142613128cf181ad8c1dc3ce214dbcdba062ce" exitCode=0 Nov 28 14:55:29 crc kubenswrapper[4817]: I1128 14:55:29.167151 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" event={"ID":"3200157e-779d-485c-803e-1b0d092c02bb","Type":"ContainerDied","Data":"4019c41d50178edb39d8d3c852142613128cf181ad8c1dc3ce214dbcdba062ce"} Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.570297 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.658652 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-inventory\") pod \"3200157e-779d-485c-803e-1b0d092c02bb\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.658879 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-ssh-key\") pod \"3200157e-779d-485c-803e-1b0d092c02bb\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.658936 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz272\" (UniqueName: \"kubernetes.io/projected/3200157e-779d-485c-803e-1b0d092c02bb-kube-api-access-zz272\") pod \"3200157e-779d-485c-803e-1b0d092c02bb\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.658989 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-bootstrap-combined-ca-bundle\") pod \"3200157e-779d-485c-803e-1b0d092c02bb\" (UID: \"3200157e-779d-485c-803e-1b0d092c02bb\") " Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.665151 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3200157e-779d-485c-803e-1b0d092c02bb-kube-api-access-zz272" (OuterVolumeSpecName: "kube-api-access-zz272") pod "3200157e-779d-485c-803e-1b0d092c02bb" (UID: "3200157e-779d-485c-803e-1b0d092c02bb"). InnerVolumeSpecName "kube-api-access-zz272". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.665520 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3200157e-779d-485c-803e-1b0d092c02bb" (UID: "3200157e-779d-485c-803e-1b0d092c02bb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.689667 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3200157e-779d-485c-803e-1b0d092c02bb" (UID: "3200157e-779d-485c-803e-1b0d092c02bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.704460 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-inventory" (OuterVolumeSpecName: "inventory") pod "3200157e-779d-485c-803e-1b0d092c02bb" (UID: "3200157e-779d-485c-803e-1b0d092c02bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.761164 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.761195 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.761205 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz272\" (UniqueName: \"kubernetes.io/projected/3200157e-779d-485c-803e-1b0d092c02bb-kube-api-access-zz272\") on node \"crc\" DevicePath \"\"" Nov 28 14:55:30 crc kubenswrapper[4817]: I1128 14:55:30.761228 4817 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3200157e-779d-485c-803e-1b0d092c02bb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.187388 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" event={"ID":"3200157e-779d-485c-803e-1b0d092c02bb","Type":"ContainerDied","Data":"d24ced0ca42b845f98727767ef44dd4ab3677d206b9e088f7d581d9e73918369"} Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.187424 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d24ced0ca42b845f98727767ef44dd4ab3677d206b9e088f7d581d9e73918369" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.187474 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.267437 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q"] Nov 28 14:55:31 crc kubenswrapper[4817]: E1128 14:55:31.268107 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3200157e-779d-485c-803e-1b0d092c02bb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.268180 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3200157e-779d-485c-803e-1b0d092c02bb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.268486 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3200157e-779d-485c-803e-1b0d092c02bb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.269304 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.273882 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.273885 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.274397 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.274578 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.277630 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q"] Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.372221 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57rjd\" (UniqueName: \"kubernetes.io/projected/0452060d-61d9-49c6-b16c-522ded9cb4e8-kube-api-access-57rjd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.372446 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.372473 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.473641 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57rjd\" (UniqueName: \"kubernetes.io/projected/0452060d-61d9-49c6-b16c-522ded9cb4e8-kube-api-access-57rjd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.473790 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.473815 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.478677 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.478823 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.500910 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57rjd\" (UniqueName: \"kubernetes.io/projected/0452060d-61d9-49c6-b16c-522ded9cb4e8-kube-api-access-57rjd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:31 crc kubenswrapper[4817]: I1128 14:55:31.653869 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:55:32 crc kubenswrapper[4817]: I1128 14:55:32.224752 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q"] Nov 28 14:55:32 crc kubenswrapper[4817]: I1128 14:55:32.226707 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:55:33 crc kubenswrapper[4817]: I1128 14:55:33.211842 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" event={"ID":"0452060d-61d9-49c6-b16c-522ded9cb4e8","Type":"ContainerStarted","Data":"340e2661b3c7f38b2366332b4ec01964b69998c71c71f1727104de38d88cf494"} Nov 28 14:55:33 crc kubenswrapper[4817]: I1128 14:55:33.212133 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" event={"ID":"0452060d-61d9-49c6-b16c-522ded9cb4e8","Type":"ContainerStarted","Data":"597e50027f4d3122feb968ef44c0db1e7a992695e1a9e10c8fd282f6704aaad5"} Nov 28 14:55:33 crc kubenswrapper[4817]: I1128 14:55:33.235155 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" podStartSLOduration=1.653789016 podStartE2EDuration="2.235129129s" podCreationTimestamp="2025-11-28 14:55:31 +0000 UTC" firstStartedPulling="2025-11-28 14:55:32.226510763 +0000 UTC m=+1554.814489029" lastFinishedPulling="2025-11-28 14:55:32.807850866 +0000 UTC m=+1555.395829142" observedRunningTime="2025-11-28 14:55:33.232593686 +0000 UTC m=+1555.820571972" watchObservedRunningTime="2025-11-28 14:55:33.235129129 +0000 UTC m=+1555.823107435" Nov 28 14:55:53 crc kubenswrapper[4817]: I1128 14:55:53.061770 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-93e8-account-create-update-g8fpv"] Nov 28 14:55:53 crc kubenswrapper[4817]: I1128 14:55:53.081886 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-sdjgh"] Nov 28 14:55:53 crc kubenswrapper[4817]: I1128 14:55:53.094449 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-sdjgh"] Nov 28 14:55:53 crc kubenswrapper[4817]: I1128 14:55:53.104950 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-93e8-account-create-update-g8fpv"] Nov 28 14:55:53 crc kubenswrapper[4817]: I1128 14:55:53.761409 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="632a370a-92b6-4f1c-beeb-c6a7a5f562ff" path="/var/lib/kubelet/pods/632a370a-92b6-4f1c-beeb-c6a7a5f562ff/volumes" Nov 28 14:55:53 crc kubenswrapper[4817]: I1128 14:55:53.763073 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b50fef50-1b7c-4f91-8f53-258516e099d7" path="/var/lib/kubelet/pods/b50fef50-1b7c-4f91-8f53-258516e099d7/volumes" Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.045173 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.045261 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.052638 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c35b-account-create-update-jwxrp"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.071179 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-bf97-account-create-update-wb4z8"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.081840 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c35b-account-create-update-jwxrp"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.092579 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-bf97-account-create-update-wb4z8"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.100610 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pvwxr"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.109208 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-q4zm8"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.117138 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-q4zm8"] Nov 28 14:55:54 crc kubenswrapper[4817]: I1128 14:55:54.124290 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pvwxr"] Nov 28 14:55:55 crc kubenswrapper[4817]: I1128 14:55:55.752034 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c60d2ea-9bc6-48d2-a82a-788bf45b2381" path="/var/lib/kubelet/pods/5c60d2ea-9bc6-48d2-a82a-788bf45b2381/volumes" Nov 28 14:55:55 crc kubenswrapper[4817]: I1128 14:55:55.753325 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c8e9ae0-c83a-4280-9d20-c03f045203d3" path="/var/lib/kubelet/pods/6c8e9ae0-c83a-4280-9d20-c03f045203d3/volumes" Nov 28 14:55:55 crc kubenswrapper[4817]: I1128 14:55:55.754098 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c6cb42-23bc-48e1-ba8a-671f454d5aac" path="/var/lib/kubelet/pods/92c6cb42-23bc-48e1-ba8a-671f454d5aac/volumes" Nov 28 14:55:55 crc kubenswrapper[4817]: I1128 14:55:55.754956 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14" path="/var/lib/kubelet/pods/eacbde1d-ce9d-460f-a9d4-e6bd7ea19c14/volumes" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.190487 4817 scope.go:117] "RemoveContainer" containerID="c0e3885d125a5380218f976fa5fb5c0d354a9c86f89898a2160cb8da22bbe3a5" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.234473 4817 scope.go:117] "RemoveContainer" containerID="dc89bf855fd55526e5c99aa2850da18a9ca6372eb5bf070a2897f42a29cc90d1" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.292677 4817 scope.go:117] "RemoveContainer" containerID="599f2eb3493ecf17c978bbf915382f7a92c014240322321f14cb98433c849039" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.325126 4817 scope.go:117] "RemoveContainer" containerID="8dadfdf37a4c2d9a4084785cf7d5fab3e8860e24e144650fa6cd148c8d468748" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.369557 4817 scope.go:117] "RemoveContainer" containerID="e563b7daf40fd96fc19785df2b1cf6c5fc535eca965c42a1c00f6e9cc00b4ce5" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.406283 4817 scope.go:117] "RemoveContainer" containerID="d2006d514f24f53332336ad7fae97618d443db30e7155b77197a16537c18dba0" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.454441 4817 scope.go:117] "RemoveContainer" containerID="4d985f2a6596a8516cc62873eeb632b2a30cfc58126fbd0d3ec15082ccc1d9bb" Nov 28 14:56:11 crc kubenswrapper[4817]: I1128 14:56:11.495971 4817 scope.go:117] "RemoveContainer" containerID="ba50ebff34de984713a13742308c1ad582ca9e8dbbc1eb26869ea57d72d815da" Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.040868 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-5p8hr"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.052566 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-5p8hr"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.062799 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-djdx9"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.081822 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b6c2-account-create-update-f4v77"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.091377 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-djdx9"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.100204 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f9d9-account-create-update-v74l9"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.108233 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-tncs8"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.116885 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-tncs8"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.123969 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b6c2-account-create-update-f4v77"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.130631 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-pbdln"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.151796 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f9d9-account-create-update-v74l9"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.162699 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-pbdln"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.171279 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-392c-account-create-update-bv6kv"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.179019 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-49a4-account-create-update-vmg9v"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.187027 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-392c-account-create-update-bv6kv"] Nov 28 14:56:20 crc kubenswrapper[4817]: I1128 14:56:20.195930 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-49a4-account-create-update-vmg9v"] Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.759318 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4763ad3a-c76f-4750-8318-d18b1831a3e8" path="/var/lib/kubelet/pods/4763ad3a-c76f-4750-8318-d18b1831a3e8/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.762380 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56eb98d9-0aff-4113-b6f0-34329d3b476d" path="/var/lib/kubelet/pods/56eb98d9-0aff-4113-b6f0-34329d3b476d/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.765420 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd26d47-662e-4943-ac9d-fae6e8f75a81" path="/var/lib/kubelet/pods/5fd26d47-662e-4943-ac9d-fae6e8f75a81/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.766614 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1e6a5f-677d-43bc-9790-9f85ade5db4d" path="/var/lib/kubelet/pods/6d1e6a5f-677d-43bc-9790-9f85ade5db4d/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.767841 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7599e834-c07e-4db9-828b-675b53df2aa6" path="/var/lib/kubelet/pods/7599e834-c07e-4db9-828b-675b53df2aa6/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.769791 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3c64713-aa61-41bf-9e97-4edc35292d40" path="/var/lib/kubelet/pods/d3c64713-aa61-41bf-9e97-4edc35292d40/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.770943 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d46f3006-1347-4670-ab14-1901562df15a" path="/var/lib/kubelet/pods/d46f3006-1347-4670-ab14-1901562df15a/volumes" Nov 28 14:56:21 crc kubenswrapper[4817]: I1128 14:56:21.772484 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9853b7-9f8c-4b0c-8b02-4bf014636556" path="/var/lib/kubelet/pods/eb9853b7-9f8c-4b0c-8b02-4bf014636556/volumes" Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.045614 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.046179 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.046247 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.047538 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.047652 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" gracePeriod=600 Nov 28 14:56:24 crc kubenswrapper[4817]: E1128 14:56:24.179891 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.794358 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" exitCode=0 Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.794410 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290"} Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.794763 4817 scope.go:117] "RemoveContainer" containerID="d7852240b73d59ba9afb5c799b3105da2bc45e3baeb4492c4a8478d7aca2dfe6" Nov 28 14:56:24 crc kubenswrapper[4817]: I1128 14:56:24.796172 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:56:24 crc kubenswrapper[4817]: E1128 14:56:24.796706 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.040841 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-5gzqs"] Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.051700 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-5gzqs"] Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.667173 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7vhvn"] Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.673467 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.684774 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vhvn"] Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.699024 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-utilities\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.699254 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d5tm\" (UniqueName: \"kubernetes.io/projected/21243ab9-36e7-4771-b1c9-3a5ff1faac28-kube-api-access-7d5tm\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.699452 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-catalog-content\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.754122 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77135fc7-4af7-4119-9153-525f2e7862c4" path="/var/lib/kubelet/pods/77135fc7-4af7-4119-9153-525f2e7862c4/volumes" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.801059 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-catalog-content\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.801191 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-utilities\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.801316 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d5tm\" (UniqueName: \"kubernetes.io/projected/21243ab9-36e7-4771-b1c9-3a5ff1faac28-kube-api-access-7d5tm\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.802924 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-catalog-content\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.803093 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-utilities\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:25 crc kubenswrapper[4817]: I1128 14:56:25.820838 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d5tm\" (UniqueName: \"kubernetes.io/projected/21243ab9-36e7-4771-b1c9-3a5ff1faac28-kube-api-access-7d5tm\") pod \"redhat-marketplace-7vhvn\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:26 crc kubenswrapper[4817]: I1128 14:56:26.029559 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:26 crc kubenswrapper[4817]: I1128 14:56:26.495475 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vhvn"] Nov 28 14:56:26 crc kubenswrapper[4817]: W1128 14:56:26.503405 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21243ab9_36e7_4771_b1c9_3a5ff1faac28.slice/crio-32417622b2e83912d0edcd71b9cbf34f0c21f4409ab136920d9d91f6a1d34033 WatchSource:0}: Error finding container 32417622b2e83912d0edcd71b9cbf34f0c21f4409ab136920d9d91f6a1d34033: Status 404 returned error can't find the container with id 32417622b2e83912d0edcd71b9cbf34f0c21f4409ab136920d9d91f6a1d34033 Nov 28 14:56:26 crc kubenswrapper[4817]: I1128 14:56:26.816412 4817 generic.go:334] "Generic (PLEG): container finished" podID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerID="6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028" exitCode=0 Nov 28 14:56:26 crc kubenswrapper[4817]: I1128 14:56:26.816481 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vhvn" event={"ID":"21243ab9-36e7-4771-b1c9-3a5ff1faac28","Type":"ContainerDied","Data":"6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028"} Nov 28 14:56:26 crc kubenswrapper[4817]: I1128 14:56:26.816664 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vhvn" event={"ID":"21243ab9-36e7-4771-b1c9-3a5ff1faac28","Type":"ContainerStarted","Data":"32417622b2e83912d0edcd71b9cbf34f0c21f4409ab136920d9d91f6a1d34033"} Nov 28 14:56:27 crc kubenswrapper[4817]: I1128 14:56:27.830969 4817 generic.go:334] "Generic (PLEG): container finished" podID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerID="f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed" exitCode=0 Nov 28 14:56:27 crc kubenswrapper[4817]: I1128 14:56:27.831041 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vhvn" event={"ID":"21243ab9-36e7-4771-b1c9-3a5ff1faac28","Type":"ContainerDied","Data":"f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed"} Nov 28 14:56:28 crc kubenswrapper[4817]: I1128 14:56:28.057908 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gzplt"] Nov 28 14:56:28 crc kubenswrapper[4817]: I1128 14:56:28.077010 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gzplt"] Nov 28 14:56:28 crc kubenswrapper[4817]: I1128 14:56:28.844909 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vhvn" event={"ID":"21243ab9-36e7-4771-b1c9-3a5ff1faac28","Type":"ContainerStarted","Data":"91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526"} Nov 28 14:56:28 crc kubenswrapper[4817]: I1128 14:56:28.869210 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7vhvn" podStartSLOduration=2.341525429 podStartE2EDuration="3.869191105s" podCreationTimestamp="2025-11-28 14:56:25 +0000 UTC" firstStartedPulling="2025-11-28 14:56:26.819219797 +0000 UTC m=+1609.407198103" lastFinishedPulling="2025-11-28 14:56:28.346885513 +0000 UTC m=+1610.934863779" observedRunningTime="2025-11-28 14:56:28.863499963 +0000 UTC m=+1611.451478239" watchObservedRunningTime="2025-11-28 14:56:28.869191105 +0000 UTC m=+1611.457169371" Nov 28 14:56:29 crc kubenswrapper[4817]: I1128 14:56:29.756215 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493a9cc0-bb5a-463c-be84-c6def176bac7" path="/var/lib/kubelet/pods/493a9cc0-bb5a-463c-be84-c6def176bac7/volumes" Nov 28 14:56:36 crc kubenswrapper[4817]: I1128 14:56:36.030104 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:36 crc kubenswrapper[4817]: I1128 14:56:36.030709 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:36 crc kubenswrapper[4817]: I1128 14:56:36.102775 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:36 crc kubenswrapper[4817]: I1128 14:56:36.736905 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:56:36 crc kubenswrapper[4817]: E1128 14:56:36.737881 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:56:36 crc kubenswrapper[4817]: I1128 14:56:36.980369 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:37 crc kubenswrapper[4817]: I1128 14:56:37.032335 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vhvn"] Nov 28 14:56:38 crc kubenswrapper[4817]: I1128 14:56:38.937445 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7vhvn" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="registry-server" containerID="cri-o://91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526" gracePeriod=2 Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.453754 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.500827 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d5tm\" (UniqueName: \"kubernetes.io/projected/21243ab9-36e7-4771-b1c9-3a5ff1faac28-kube-api-access-7d5tm\") pod \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.501028 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-catalog-content\") pod \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.501087 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-utilities\") pod \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\" (UID: \"21243ab9-36e7-4771-b1c9-3a5ff1faac28\") " Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.502256 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-utilities" (OuterVolumeSpecName: "utilities") pod "21243ab9-36e7-4771-b1c9-3a5ff1faac28" (UID: "21243ab9-36e7-4771-b1c9-3a5ff1faac28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.506584 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21243ab9-36e7-4771-b1c9-3a5ff1faac28-kube-api-access-7d5tm" (OuterVolumeSpecName: "kube-api-access-7d5tm") pod "21243ab9-36e7-4771-b1c9-3a5ff1faac28" (UID: "21243ab9-36e7-4771-b1c9-3a5ff1faac28"). InnerVolumeSpecName "kube-api-access-7d5tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.522248 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21243ab9-36e7-4771-b1c9-3a5ff1faac28" (UID: "21243ab9-36e7-4771-b1c9-3a5ff1faac28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.604672 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.604711 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d5tm\" (UniqueName: \"kubernetes.io/projected/21243ab9-36e7-4771-b1c9-3a5ff1faac28-kube-api-access-7d5tm\") on node \"crc\" DevicePath \"\"" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.604737 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21243ab9-36e7-4771-b1c9-3a5ff1faac28-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.961467 4817 generic.go:334] "Generic (PLEG): container finished" podID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerID="91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526" exitCode=0 Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.961897 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vhvn" event={"ID":"21243ab9-36e7-4771-b1c9-3a5ff1faac28","Type":"ContainerDied","Data":"91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526"} Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.961937 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vhvn" event={"ID":"21243ab9-36e7-4771-b1c9-3a5ff1faac28","Type":"ContainerDied","Data":"32417622b2e83912d0edcd71b9cbf34f0c21f4409ab136920d9d91f6a1d34033"} Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.961962 4817 scope.go:117] "RemoveContainer" containerID="91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526" Nov 28 14:56:39 crc kubenswrapper[4817]: I1128 14:56:39.962159 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vhvn" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.016172 4817 scope.go:117] "RemoveContainer" containerID="f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.016453 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vhvn"] Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.024976 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vhvn"] Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.070296 4817 scope.go:117] "RemoveContainer" containerID="6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.127813 4817 scope.go:117] "RemoveContainer" containerID="91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526" Nov 28 14:56:40 crc kubenswrapper[4817]: E1128 14:56:40.130824 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526\": container with ID starting with 91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526 not found: ID does not exist" containerID="91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.130857 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526"} err="failed to get container status \"91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526\": rpc error: code = NotFound desc = could not find container \"91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526\": container with ID starting with 91277d6e06cf1bc1731bf369eddc3ca44d74ae10092cdf4b9f02d5e00c730526 not found: ID does not exist" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.130878 4817 scope.go:117] "RemoveContainer" containerID="f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed" Nov 28 14:56:40 crc kubenswrapper[4817]: E1128 14:56:40.131243 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed\": container with ID starting with f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed not found: ID does not exist" containerID="f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.131274 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed"} err="failed to get container status \"f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed\": rpc error: code = NotFound desc = could not find container \"f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed\": container with ID starting with f1ccbabc712ffee4a070062df2036f99404996e439af77a4992729c4bcee66ed not found: ID does not exist" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.131293 4817 scope.go:117] "RemoveContainer" containerID="6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028" Nov 28 14:56:40 crc kubenswrapper[4817]: E1128 14:56:40.131567 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028\": container with ID starting with 6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028 not found: ID does not exist" containerID="6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028" Nov 28 14:56:40 crc kubenswrapper[4817]: I1128 14:56:40.131593 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028"} err="failed to get container status \"6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028\": rpc error: code = NotFound desc = could not find container \"6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028\": container with ID starting with 6d165efa9c9eaaa7651d2aecc33bd7868ba40d9a9de29f473d2173135010f028 not found: ID does not exist" Nov 28 14:56:41 crc kubenswrapper[4817]: I1128 14:56:41.746022 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" path="/var/lib/kubelet/pods/21243ab9-36e7-4771-b1c9-3a5ff1faac28/volumes" Nov 28 14:56:47 crc kubenswrapper[4817]: I1128 14:56:47.745184 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:56:47 crc kubenswrapper[4817]: E1128 14:56:47.746938 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.674663 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6gct5"] Nov 28 14:57:02 crc kubenswrapper[4817]: E1128 14:57:02.677568 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="extract-content" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.677604 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="extract-content" Nov 28 14:57:02 crc kubenswrapper[4817]: E1128 14:57:02.677643 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="registry-server" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.677660 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="registry-server" Nov 28 14:57:02 crc kubenswrapper[4817]: E1128 14:57:02.677748 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="extract-utilities" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.677765 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="extract-utilities" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.679377 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="21243ab9-36e7-4771-b1c9-3a5ff1faac28" containerName="registry-server" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.690287 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.715256 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gct5"] Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.737702 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:57:02 crc kubenswrapper[4817]: E1128 14:57:02.738253 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.868418 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-utilities\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.868468 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-catalog-content\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.868525 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82rnh\" (UniqueName: \"kubernetes.io/projected/bf619100-a35d-4e77-bca2-f6d741688668-kube-api-access-82rnh\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.972509 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-utilities\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.972583 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-catalog-content\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.972631 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82rnh\" (UniqueName: \"kubernetes.io/projected/bf619100-a35d-4e77-bca2-f6d741688668-kube-api-access-82rnh\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.973340 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-utilities\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.973483 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-catalog-content\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:02 crc kubenswrapper[4817]: I1128 14:57:02.993559 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82rnh\" (UniqueName: \"kubernetes.io/projected/bf619100-a35d-4e77-bca2-f6d741688668-kube-api-access-82rnh\") pod \"certified-operators-6gct5\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:03 crc kubenswrapper[4817]: I1128 14:57:03.040128 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:03 crc kubenswrapper[4817]: I1128 14:57:03.568104 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gct5"] Nov 28 14:57:04 crc kubenswrapper[4817]: I1128 14:57:04.243871 4817 generic.go:334] "Generic (PLEG): container finished" podID="bf619100-a35d-4e77-bca2-f6d741688668" containerID="41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53" exitCode=0 Nov 28 14:57:04 crc kubenswrapper[4817]: I1128 14:57:04.243944 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerDied","Data":"41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53"} Nov 28 14:57:04 crc kubenswrapper[4817]: I1128 14:57:04.244182 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerStarted","Data":"5a0020344bc7ee9de888ef816cb5e3a2b674a8bdcd9cfe077c9ef3911fdca3e1"} Nov 28 14:57:05 crc kubenswrapper[4817]: I1128 14:57:05.255931 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerStarted","Data":"05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf"} Nov 28 14:57:06 crc kubenswrapper[4817]: I1128 14:57:06.271068 4817 generic.go:334] "Generic (PLEG): container finished" podID="bf619100-a35d-4e77-bca2-f6d741688668" containerID="05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf" exitCode=0 Nov 28 14:57:06 crc kubenswrapper[4817]: I1128 14:57:06.271151 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerDied","Data":"05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf"} Nov 28 14:57:07 crc kubenswrapper[4817]: I1128 14:57:07.292990 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerStarted","Data":"a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a"} Nov 28 14:57:07 crc kubenswrapper[4817]: I1128 14:57:07.321546 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6gct5" podStartSLOduration=2.712383892 podStartE2EDuration="5.321531842s" podCreationTimestamp="2025-11-28 14:57:02 +0000 UTC" firstStartedPulling="2025-11-28 14:57:04.246541638 +0000 UTC m=+1646.834519944" lastFinishedPulling="2025-11-28 14:57:06.855689618 +0000 UTC m=+1649.443667894" observedRunningTime="2025-11-28 14:57:07.318584878 +0000 UTC m=+1649.906563184" watchObservedRunningTime="2025-11-28 14:57:07.321531842 +0000 UTC m=+1649.909510108" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.685919 4817 scope.go:117] "RemoveContainer" containerID="8ef295d27a65648074178bfa9277675404a7fa65b59f04a5fbb37733f2790a0d" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.739921 4817 scope.go:117] "RemoveContainer" containerID="4c9c61a9d17a53afd59ecb4c1ba414ab39925af267cb37f8992436ba5b1b16a6" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.780928 4817 scope.go:117] "RemoveContainer" containerID="6deb01ae50ab392790d63376f8ce2bea35d7eb657a4953fedcc7edd7af4eb0f6" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.818188 4817 scope.go:117] "RemoveContainer" containerID="aff87f904424add4328aa2d941ae88c705fa26e457ebbf6614ebb965b4663553" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.863864 4817 scope.go:117] "RemoveContainer" containerID="6794b42950e35f9c2253282e6bf0a236d3a1ae64841d88538063f849c4f020fc" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.903561 4817 scope.go:117] "RemoveContainer" containerID="37d82d58bc3090b9566d85a315d65786e873485708b662958686d1bdb46af9d9" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.970529 4817 scope.go:117] "RemoveContainer" containerID="1694bf4ac9a28943f0d84e487884ffe418ebf912fbe4706753797183e070c064" Nov 28 14:57:11 crc kubenswrapper[4817]: I1128 14:57:11.995412 4817 scope.go:117] "RemoveContainer" containerID="1b312db9d53b12e3fe3eadd74e468e16d43c3db1cf8d0d61229e80863e2a8ac1" Nov 28 14:57:12 crc kubenswrapper[4817]: I1128 14:57:12.018152 4817 scope.go:117] "RemoveContainer" containerID="d0d85161488f3692438f96efeda13be57155cde41e7bb429925c0ff295ef0da7" Nov 28 14:57:12 crc kubenswrapper[4817]: I1128 14:57:12.041421 4817 scope.go:117] "RemoveContainer" containerID="b7dfc22974a5c6af3004cfcd83ea18cce90abbb2e1517132669651b7eba7f60b" Nov 28 14:57:13 crc kubenswrapper[4817]: I1128 14:57:13.041218 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:13 crc kubenswrapper[4817]: I1128 14:57:13.041645 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:13 crc kubenswrapper[4817]: I1128 14:57:13.111582 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:13 crc kubenswrapper[4817]: I1128 14:57:13.423138 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:13 crc kubenswrapper[4817]: I1128 14:57:13.473962 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gct5"] Nov 28 14:57:14 crc kubenswrapper[4817]: I1128 14:57:14.737500 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:57:14 crc kubenswrapper[4817]: E1128 14:57:14.738175 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.374046 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6gct5" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="registry-server" containerID="cri-o://a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a" gracePeriod=2 Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.858355 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.945981 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-catalog-content\") pod \"bf619100-a35d-4e77-bca2-f6d741688668\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.946324 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-utilities\") pod \"bf619100-a35d-4e77-bca2-f6d741688668\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.946415 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82rnh\" (UniqueName: \"kubernetes.io/projected/bf619100-a35d-4e77-bca2-f6d741688668-kube-api-access-82rnh\") pod \"bf619100-a35d-4e77-bca2-f6d741688668\" (UID: \"bf619100-a35d-4e77-bca2-f6d741688668\") " Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.947764 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-utilities" (OuterVolumeSpecName: "utilities") pod "bf619100-a35d-4e77-bca2-f6d741688668" (UID: "bf619100-a35d-4e77-bca2-f6d741688668"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.952716 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf619100-a35d-4e77-bca2-f6d741688668-kube-api-access-82rnh" (OuterVolumeSpecName: "kube-api-access-82rnh") pod "bf619100-a35d-4e77-bca2-f6d741688668" (UID: "bf619100-a35d-4e77-bca2-f6d741688668"). InnerVolumeSpecName "kube-api-access-82rnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:57:15 crc kubenswrapper[4817]: I1128 14:57:15.997640 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf619100-a35d-4e77-bca2-f6d741688668" (UID: "bf619100-a35d-4e77-bca2-f6d741688668"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.048891 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.048939 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82rnh\" (UniqueName: \"kubernetes.io/projected/bf619100-a35d-4e77-bca2-f6d741688668-kube-api-access-82rnh\") on node \"crc\" DevicePath \"\"" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.048957 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf619100-a35d-4e77-bca2-f6d741688668-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.390724 4817 generic.go:334] "Generic (PLEG): container finished" podID="bf619100-a35d-4e77-bca2-f6d741688668" containerID="a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a" exitCode=0 Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.390843 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerDied","Data":"a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a"} Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.390895 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gct5" event={"ID":"bf619100-a35d-4e77-bca2-f6d741688668","Type":"ContainerDied","Data":"5a0020344bc7ee9de888ef816cb5e3a2b674a8bdcd9cfe077c9ef3911fdca3e1"} Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.390916 4817 scope.go:117] "RemoveContainer" containerID="a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.390842 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gct5" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.430251 4817 scope.go:117] "RemoveContainer" containerID="05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.438293 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gct5"] Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.449027 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6gct5"] Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.455890 4817 scope.go:117] "RemoveContainer" containerID="41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.501924 4817 scope.go:117] "RemoveContainer" containerID="a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a" Nov 28 14:57:16 crc kubenswrapper[4817]: E1128 14:57:16.502493 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a\": container with ID starting with a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a not found: ID does not exist" containerID="a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.502529 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a"} err="failed to get container status \"a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a\": rpc error: code = NotFound desc = could not find container \"a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a\": container with ID starting with a0a46af2f0d17b6d887c736a43aa8ce87d6ad6a1aaf26cebdc0a9a0b148d982a not found: ID does not exist" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.502555 4817 scope.go:117] "RemoveContainer" containerID="05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf" Nov 28 14:57:16 crc kubenswrapper[4817]: E1128 14:57:16.502995 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf\": container with ID starting with 05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf not found: ID does not exist" containerID="05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.503096 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf"} err="failed to get container status \"05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf\": rpc error: code = NotFound desc = could not find container \"05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf\": container with ID starting with 05ffbf089a4bfe76263e79b1a0868ba87edcdb1814a348d2e6a404f1c80806bf not found: ID does not exist" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.503207 4817 scope.go:117] "RemoveContainer" containerID="41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53" Nov 28 14:57:16 crc kubenswrapper[4817]: E1128 14:57:16.503651 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53\": container with ID starting with 41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53 not found: ID does not exist" containerID="41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53" Nov 28 14:57:16 crc kubenswrapper[4817]: I1128 14:57:16.503691 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53"} err="failed to get container status \"41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53\": rpc error: code = NotFound desc = could not find container \"41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53\": container with ID starting with 41e5d23c3b0cbba51bc93f924fe7702f26755bcef60619f77c8d5cb90e0f2b53 not found: ID does not exist" Nov 28 14:57:17 crc kubenswrapper[4817]: I1128 14:57:17.753383 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf619100-a35d-4e77-bca2-f6d741688668" path="/var/lib/kubelet/pods/bf619100-a35d-4e77-bca2-f6d741688668/volumes" Nov 28 14:57:22 crc kubenswrapper[4817]: I1128 14:57:22.057415 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gvbcw"] Nov 28 14:57:22 crc kubenswrapper[4817]: I1128 14:57:22.073328 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gvbcw"] Nov 28 14:57:22 crc kubenswrapper[4817]: I1128 14:57:22.082495 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-lgl42"] Nov 28 14:57:22 crc kubenswrapper[4817]: I1128 14:57:22.089828 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-lgl42"] Nov 28 14:57:23 crc kubenswrapper[4817]: I1128 14:57:23.751182 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d26f30b-efdc-40ef-9ae5-ca78e79aba41" path="/var/lib/kubelet/pods/3d26f30b-efdc-40ef-9ae5-ca78e79aba41/volumes" Nov 28 14:57:23 crc kubenswrapper[4817]: I1128 14:57:23.752164 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4482a16d-c783-4609-9db4-4e9eb3cb3c23" path="/var/lib/kubelet/pods/4482a16d-c783-4609-9db4-4e9eb3cb3c23/volumes" Nov 28 14:57:26 crc kubenswrapper[4817]: I1128 14:57:26.056823 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-fbfcr"] Nov 28 14:57:26 crc kubenswrapper[4817]: I1128 14:57:26.067559 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-fbfcr"] Nov 28 14:57:26 crc kubenswrapper[4817]: I1128 14:57:26.737574 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:57:26 crc kubenswrapper[4817]: E1128 14:57:26.738138 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:57:27 crc kubenswrapper[4817]: I1128 14:57:27.028593 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-fflrd"] Nov 28 14:57:27 crc kubenswrapper[4817]: I1128 14:57:27.035687 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-fflrd"] Nov 28 14:57:27 crc kubenswrapper[4817]: I1128 14:57:27.765475 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d2d10b9-2269-4654-a979-cf2c9a8c65dc" path="/var/lib/kubelet/pods/5d2d10b9-2269-4654-a979-cf2c9a8c65dc/volumes" Nov 28 14:57:27 crc kubenswrapper[4817]: I1128 14:57:27.767065 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da990a2-1382-4707-8311-5dd46e4f537c" path="/var/lib/kubelet/pods/6da990a2-1382-4707-8311-5dd46e4f537c/volumes" Nov 28 14:57:32 crc kubenswrapper[4817]: I1128 14:57:32.038075 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-5q2qk"] Nov 28 14:57:32 crc kubenswrapper[4817]: I1128 14:57:32.054247 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-5q2qk"] Nov 28 14:57:33 crc kubenswrapper[4817]: I1128 14:57:33.028895 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-mpzzx"] Nov 28 14:57:33 crc kubenswrapper[4817]: I1128 14:57:33.037243 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-mpzzx"] Nov 28 14:57:33 crc kubenswrapper[4817]: I1128 14:57:33.754584 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11e39ba1-49a5-4289-a913-10b87e82f4e0" path="/var/lib/kubelet/pods/11e39ba1-49a5-4289-a913-10b87e82f4e0/volumes" Nov 28 14:57:33 crc kubenswrapper[4817]: I1128 14:57:33.755836 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca79eb0e-3393-4730-ba3a-c56fc3c6623f" path="/var/lib/kubelet/pods/ca79eb0e-3393-4730-ba3a-c56fc3c6623f/volumes" Nov 28 14:57:37 crc kubenswrapper[4817]: I1128 14:57:37.764712 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:57:37 crc kubenswrapper[4817]: E1128 14:57:37.766004 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:57:38 crc kubenswrapper[4817]: I1128 14:57:38.640572 4817 generic.go:334] "Generic (PLEG): container finished" podID="0452060d-61d9-49c6-b16c-522ded9cb4e8" containerID="340e2661b3c7f38b2366332b4ec01964b69998c71c71f1727104de38d88cf494" exitCode=0 Nov 28 14:57:38 crc kubenswrapper[4817]: I1128 14:57:38.641006 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" event={"ID":"0452060d-61d9-49c6-b16c-522ded9cb4e8","Type":"ContainerDied","Data":"340e2661b3c7f38b2366332b4ec01964b69998c71c71f1727104de38d88cf494"} Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.130470 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.226200 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-ssh-key\") pod \"0452060d-61d9-49c6-b16c-522ded9cb4e8\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.226334 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57rjd\" (UniqueName: \"kubernetes.io/projected/0452060d-61d9-49c6-b16c-522ded9cb4e8-kube-api-access-57rjd\") pod \"0452060d-61d9-49c6-b16c-522ded9cb4e8\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.226459 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-inventory\") pod \"0452060d-61d9-49c6-b16c-522ded9cb4e8\" (UID: \"0452060d-61d9-49c6-b16c-522ded9cb4e8\") " Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.233253 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0452060d-61d9-49c6-b16c-522ded9cb4e8-kube-api-access-57rjd" (OuterVolumeSpecName: "kube-api-access-57rjd") pod "0452060d-61d9-49c6-b16c-522ded9cb4e8" (UID: "0452060d-61d9-49c6-b16c-522ded9cb4e8"). InnerVolumeSpecName "kube-api-access-57rjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.253446 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0452060d-61d9-49c6-b16c-522ded9cb4e8" (UID: "0452060d-61d9-49c6-b16c-522ded9cb4e8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.256798 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-inventory" (OuterVolumeSpecName: "inventory") pod "0452060d-61d9-49c6-b16c-522ded9cb4e8" (UID: "0452060d-61d9-49c6-b16c-522ded9cb4e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.328700 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.328802 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0452060d-61d9-49c6-b16c-522ded9cb4e8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.328821 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57rjd\" (UniqueName: \"kubernetes.io/projected/0452060d-61d9-49c6-b16c-522ded9cb4e8-kube-api-access-57rjd\") on node \"crc\" DevicePath \"\"" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.683129 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.683139 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q" event={"ID":"0452060d-61d9-49c6-b16c-522ded9cb4e8","Type":"ContainerDied","Data":"597e50027f4d3122feb968ef44c0db1e7a992695e1a9e10c8fd282f6704aaad5"} Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.683600 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="597e50027f4d3122feb968ef44c0db1e7a992695e1a9e10c8fd282f6704aaad5" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.781226 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5"] Nov 28 14:57:40 crc kubenswrapper[4817]: E1128 14:57:40.781672 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0452060d-61d9-49c6-b16c-522ded9cb4e8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.781692 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0452060d-61d9-49c6-b16c-522ded9cb4e8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 28 14:57:40 crc kubenswrapper[4817]: E1128 14:57:40.781767 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="extract-utilities" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.781775 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="extract-utilities" Nov 28 14:57:40 crc kubenswrapper[4817]: E1128 14:57:40.781784 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="registry-server" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.781792 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="registry-server" Nov 28 14:57:40 crc kubenswrapper[4817]: E1128 14:57:40.781821 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="extract-content" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.781829 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="extract-content" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.782034 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0452060d-61d9-49c6-b16c-522ded9cb4e8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.782059 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf619100-a35d-4e77-bca2-f6d741688668" containerName="registry-server" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.786973 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.791786 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.792116 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.792375 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.797477 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.810990 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5"] Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.938756 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58fpd\" (UniqueName: \"kubernetes.io/projected/18ae828a-395c-4300-9460-a063a0322335-kube-api-access-58fpd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.938908 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:40 crc kubenswrapper[4817]: I1128 14:57:40.939028 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.040180 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.040303 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58fpd\" (UniqueName: \"kubernetes.io/projected/18ae828a-395c-4300-9460-a063a0322335-kube-api-access-58fpd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.040383 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.045516 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.045630 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.077373 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58fpd\" (UniqueName: \"kubernetes.io/projected/18ae828a-395c-4300-9460-a063a0322335-kube-api-access-58fpd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.153116 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:57:41 crc kubenswrapper[4817]: I1128 14:57:41.721696 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5"] Nov 28 14:57:42 crc kubenswrapper[4817]: I1128 14:57:42.725726 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" event={"ID":"18ae828a-395c-4300-9460-a063a0322335","Type":"ContainerStarted","Data":"b90829f97c1d6c830b2887dd857fc201c27d852f1bb063d95d55400c6d0a732d"} Nov 28 14:57:43 crc kubenswrapper[4817]: I1128 14:57:43.757836 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" event={"ID":"18ae828a-395c-4300-9460-a063a0322335","Type":"ContainerStarted","Data":"8f4e92699096fc42ec27721b9ea294c7028218e6c0e1e737bd77989a861592f8"} Nov 28 14:57:43 crc kubenswrapper[4817]: I1128 14:57:43.759461 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" podStartSLOduration=2.852639818 podStartE2EDuration="3.759444266s" podCreationTimestamp="2025-11-28 14:57:40 +0000 UTC" firstStartedPulling="2025-11-28 14:57:41.722299287 +0000 UTC m=+1684.310277563" lastFinishedPulling="2025-11-28 14:57:42.629103705 +0000 UTC m=+1685.217082011" observedRunningTime="2025-11-28 14:57:43.757443036 +0000 UTC m=+1686.345421332" watchObservedRunningTime="2025-11-28 14:57:43.759444266 +0000 UTC m=+1686.347422532" Nov 28 14:57:51 crc kubenswrapper[4817]: I1128 14:57:51.737260 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:57:51 crc kubenswrapper[4817]: E1128 14:57:51.738165 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:58:03 crc kubenswrapper[4817]: I1128 14:58:03.737948 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:58:03 crc kubenswrapper[4817]: E1128 14:58:03.740563 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.049022 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-16d1-account-create-update-krscn"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.056847 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-n67gd"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.064054 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-62dsg"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.077072 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-16d1-account-create-update-krscn"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.085259 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-vlrzf"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.093281 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-063a-account-create-update-95fpn"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.101648 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d5f5-account-create-update-2hjmb"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.109193 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-n67gd"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.116743 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-62dsg"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.124386 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-063a-account-create-update-95fpn"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.132127 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d5f5-account-create-update-2hjmb"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.139565 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-vlrzf"] Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.247722 4817 scope.go:117] "RemoveContainer" containerID="91fb5eccf01242cd19f09770332a91907ca62745c86c2a4555cfbaea57beb469" Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.282901 4817 scope.go:117] "RemoveContainer" containerID="a9dce10085e4a2031130e9f98ffac108e9faf70fde37937df7241bc8a088d87d" Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.350490 4817 scope.go:117] "RemoveContainer" containerID="aa36b76a97ce87e35668be6b4c9bb5840bf2a30d6a66424da3c73491b117c116" Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.389011 4817 scope.go:117] "RemoveContainer" containerID="1da30229a0fbcadc4d0e4522431651004dbe41e73c325d9cfbbf4234aaf9f99e" Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.419106 4817 scope.go:117] "RemoveContainer" containerID="4a042097560c269170cec7774bc7e066e6dfe0f4584a1b0e56eb100875839b04" Nov 28 14:58:12 crc kubenswrapper[4817]: I1128 14:58:12.476913 4817 scope.go:117] "RemoveContainer" containerID="3a4db1f3b8d1fbb685dbb132cae457f5c5e1c0600980e7fb24b0d52bea7cff61" Nov 28 14:58:13 crc kubenswrapper[4817]: I1128 14:58:13.757463 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f4757cf-18ca-406d-92ae-d975f16cb209" path="/var/lib/kubelet/pods/5f4757cf-18ca-406d-92ae-d975f16cb209/volumes" Nov 28 14:58:13 crc kubenswrapper[4817]: I1128 14:58:13.759544 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6afd8c02-96a9-44c5-a773-a137ab3df6e9" path="/var/lib/kubelet/pods/6afd8c02-96a9-44c5-a773-a137ab3df6e9/volumes" Nov 28 14:58:13 crc kubenswrapper[4817]: I1128 14:58:13.761137 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74102797-fa18-4bb1-a981-c5c359833031" path="/var/lib/kubelet/pods/74102797-fa18-4bb1-a981-c5c359833031/volumes" Nov 28 14:58:13 crc kubenswrapper[4817]: I1128 14:58:13.762539 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fe692be-892a-46b0-b168-fa66ffce6762" path="/var/lib/kubelet/pods/9fe692be-892a-46b0-b168-fa66ffce6762/volumes" Nov 28 14:58:13 crc kubenswrapper[4817]: I1128 14:58:13.765427 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f2262a-8a40-47de-9e14-d29ead210c3e" path="/var/lib/kubelet/pods/a3f2262a-8a40-47de-9e14-d29ead210c3e/volumes" Nov 28 14:58:13 crc kubenswrapper[4817]: I1128 14:58:13.766705 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1324b91-71af-44bd-9658-7727c10193c6" path="/var/lib/kubelet/pods/d1324b91-71af-44bd-9658-7727c10193c6/volumes" Nov 28 14:58:18 crc kubenswrapper[4817]: I1128 14:58:18.737600 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:58:18 crc kubenswrapper[4817]: E1128 14:58:18.738626 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:58:29 crc kubenswrapper[4817]: I1128 14:58:29.737631 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:58:29 crc kubenswrapper[4817]: E1128 14:58:29.738344 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:58:43 crc kubenswrapper[4817]: I1128 14:58:43.759201 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:58:43 crc kubenswrapper[4817]: E1128 14:58:43.760861 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:58:57 crc kubenswrapper[4817]: I1128 14:58:57.745103 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:58:57 crc kubenswrapper[4817]: E1128 14:58:57.746090 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.393858 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jzm8w"] Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.396578 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.407712 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzm8w"] Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.568083 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-catalog-content\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.568162 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-utilities\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.568487 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8xrw\" (UniqueName: \"kubernetes.io/projected/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-kube-api-access-q8xrw\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.669853 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-catalog-content\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.669918 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-utilities\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.670006 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8xrw\" (UniqueName: \"kubernetes.io/projected/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-kube-api-access-q8xrw\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.670611 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-catalog-content\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.670677 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-utilities\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.689235 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8xrw\" (UniqueName: \"kubernetes.io/projected/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-kube-api-access-q8xrw\") pod \"redhat-operators-jzm8w\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:01 crc kubenswrapper[4817]: I1128 14:59:01.715184 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:02 crc kubenswrapper[4817]: I1128 14:59:02.251410 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzm8w"] Nov 28 14:59:02 crc kubenswrapper[4817]: I1128 14:59:02.493133 4817 generic.go:334] "Generic (PLEG): container finished" podID="18ae828a-395c-4300-9460-a063a0322335" containerID="8f4e92699096fc42ec27721b9ea294c7028218e6c0e1e737bd77989a861592f8" exitCode=0 Nov 28 14:59:02 crc kubenswrapper[4817]: I1128 14:59:02.493201 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" event={"ID":"18ae828a-395c-4300-9460-a063a0322335","Type":"ContainerDied","Data":"8f4e92699096fc42ec27721b9ea294c7028218e6c0e1e737bd77989a861592f8"} Nov 28 14:59:02 crc kubenswrapper[4817]: I1128 14:59:02.495957 4817 generic.go:334] "Generic (PLEG): container finished" podID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerID="6f9242d083e6cb9425de5393c80785e1822c505bd7489e193c587b902ca3a50d" exitCode=0 Nov 28 14:59:02 crc kubenswrapper[4817]: I1128 14:59:02.495987 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerDied","Data":"6f9242d083e6cb9425de5393c80785e1822c505bd7489e193c587b902ca3a50d"} Nov 28 14:59:02 crc kubenswrapper[4817]: I1128 14:59:02.496007 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerStarted","Data":"9f57f1714f0895169fa02d6f28b6ecacd8ad6972986e1e6ff98d753350cba77a"} Nov 28 14:59:03 crc kubenswrapper[4817]: I1128 14:59:03.062160 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-68nks"] Nov 28 14:59:03 crc kubenswrapper[4817]: I1128 14:59:03.073599 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-68nks"] Nov 28 14:59:03 crc kubenswrapper[4817]: I1128 14:59:03.506495 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerStarted","Data":"c1927cff2f7c8deaf1e952bacc31a6abf309bec781ad33059e72ca6edd8b6a09"} Nov 28 14:59:03 crc kubenswrapper[4817]: I1128 14:59:03.750393 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7cf2d7d-acea-4cc4-bff9-a216dd46be04" path="/var/lib/kubelet/pods/f7cf2d7d-acea-4cc4-bff9-a216dd46be04/volumes" Nov 28 14:59:03 crc kubenswrapper[4817]: I1128 14:59:03.879876 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.043025 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-inventory\") pod \"18ae828a-395c-4300-9460-a063a0322335\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.043199 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58fpd\" (UniqueName: \"kubernetes.io/projected/18ae828a-395c-4300-9460-a063a0322335-kube-api-access-58fpd\") pod \"18ae828a-395c-4300-9460-a063a0322335\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.043335 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-ssh-key\") pod \"18ae828a-395c-4300-9460-a063a0322335\" (UID: \"18ae828a-395c-4300-9460-a063a0322335\") " Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.050890 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ae828a-395c-4300-9460-a063a0322335-kube-api-access-58fpd" (OuterVolumeSpecName: "kube-api-access-58fpd") pod "18ae828a-395c-4300-9460-a063a0322335" (UID: "18ae828a-395c-4300-9460-a063a0322335"). InnerVolumeSpecName "kube-api-access-58fpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.076499 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-inventory" (OuterVolumeSpecName: "inventory") pod "18ae828a-395c-4300-9460-a063a0322335" (UID: "18ae828a-395c-4300-9460-a063a0322335"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.079630 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18ae828a-395c-4300-9460-a063a0322335" (UID: "18ae828a-395c-4300-9460-a063a0322335"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.146435 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58fpd\" (UniqueName: \"kubernetes.io/projected/18ae828a-395c-4300-9460-a063a0322335-kube-api-access-58fpd\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.146530 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.146545 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18ae828a-395c-4300-9460-a063a0322335-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.520928 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.520848 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5" event={"ID":"18ae828a-395c-4300-9460-a063a0322335","Type":"ContainerDied","Data":"b90829f97c1d6c830b2887dd857fc201c27d852f1bb063d95d55400c6d0a732d"} Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.522920 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b90829f97c1d6c830b2887dd857fc201c27d852f1bb063d95d55400c6d0a732d" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.524536 4817 generic.go:334] "Generic (PLEG): container finished" podID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerID="c1927cff2f7c8deaf1e952bacc31a6abf309bec781ad33059e72ca6edd8b6a09" exitCode=0 Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.524589 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerDied","Data":"c1927cff2f7c8deaf1e952bacc31a6abf309bec781ad33059e72ca6edd8b6a09"} Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.622893 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8"] Nov 28 14:59:04 crc kubenswrapper[4817]: E1128 14:59:04.623309 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ae828a-395c-4300-9460-a063a0322335" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.623327 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ae828a-395c-4300-9460-a063a0322335" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.623502 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ae828a-395c-4300-9460-a063a0322335" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.624139 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.626737 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.626997 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.627316 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.627488 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.634136 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8"] Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.778256 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cdl6\" (UniqueName: \"kubernetes.io/projected/aa1fe5c5-0d23-47fb-990d-480cd18970b8-kube-api-access-9cdl6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.778488 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.778607 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.880657 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.880766 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.880843 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cdl6\" (UniqueName: \"kubernetes.io/projected/aa1fe5c5-0d23-47fb-990d-480cd18970b8-kube-api-access-9cdl6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.887120 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.902426 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.907136 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cdl6\" (UniqueName: \"kubernetes.io/projected/aa1fe5c5-0d23-47fb-990d-480cd18970b8-kube-api-access-9cdl6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:04 crc kubenswrapper[4817]: I1128 14:59:04.948032 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:05 crc kubenswrapper[4817]: I1128 14:59:05.283111 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8"] Nov 28 14:59:05 crc kubenswrapper[4817]: I1128 14:59:05.535462 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" event={"ID":"aa1fe5c5-0d23-47fb-990d-480cd18970b8","Type":"ContainerStarted","Data":"35d803e81da7a524330d0d11f676b1413ca44b9f0e4c695f4623b8412867f56d"} Nov 28 14:59:08 crc kubenswrapper[4817]: I1128 14:59:08.572573 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" event={"ID":"aa1fe5c5-0d23-47fb-990d-480cd18970b8","Type":"ContainerStarted","Data":"0473d4cd029e1ca139499dfa8b3eaaf6a1ea049651adb4c5aaf35cc9d1ad4f22"} Nov 28 14:59:08 crc kubenswrapper[4817]: I1128 14:59:08.575929 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerStarted","Data":"149480f89d52150c634752a17b8cff67e6f8e7362e7e6bc915a941dbeac628d7"} Nov 28 14:59:08 crc kubenswrapper[4817]: I1128 14:59:08.593684 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" podStartSLOduration=2.450868879 podStartE2EDuration="4.593660873s" podCreationTimestamp="2025-11-28 14:59:04 +0000 UTC" firstStartedPulling="2025-11-28 14:59:05.285138027 +0000 UTC m=+1767.873116293" lastFinishedPulling="2025-11-28 14:59:07.427930021 +0000 UTC m=+1770.015908287" observedRunningTime="2025-11-28 14:59:08.589340016 +0000 UTC m=+1771.177318292" watchObservedRunningTime="2025-11-28 14:59:08.593660873 +0000 UTC m=+1771.181639149" Nov 28 14:59:08 crc kubenswrapper[4817]: I1128 14:59:08.615342 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jzm8w" podStartSLOduration=2.7206698620000003 podStartE2EDuration="7.615321823s" podCreationTimestamp="2025-11-28 14:59:01 +0000 UTC" firstStartedPulling="2025-11-28 14:59:02.497868977 +0000 UTC m=+1765.085847243" lastFinishedPulling="2025-11-28 14:59:07.392520918 +0000 UTC m=+1769.980499204" observedRunningTime="2025-11-28 14:59:08.613689373 +0000 UTC m=+1771.201667639" watchObservedRunningTime="2025-11-28 14:59:08.615321823 +0000 UTC m=+1771.203300099" Nov 28 14:59:10 crc kubenswrapper[4817]: I1128 14:59:10.736842 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:59:10 crc kubenswrapper[4817]: E1128 14:59:10.737377 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:59:11 crc kubenswrapper[4817]: I1128 14:59:11.716121 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:11 crc kubenswrapper[4817]: I1128 14:59:11.716704 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.622540 4817 scope.go:117] "RemoveContainer" containerID="b1a5a935c61b98455ac4ca92ef26ac7fc0a899197d1b85ed577e0df4713a3cde" Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.658697 4817 scope.go:117] "RemoveContainer" containerID="dba2fd87d6cef90f93fb264da5f5255b013e0b1135f974b54190c0527dafe6aa" Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.710565 4817 scope.go:117] "RemoveContainer" containerID="e484f11b74131240b7c6f0967c8b9d2cad42bbbe3950594e6af19b79cb323028" Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.759297 4817 scope.go:117] "RemoveContainer" containerID="c212fe10c7fd47f48512470d10d45586c44e5d0bb639d9d009663532bb88c14a" Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.761844 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzm8w" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="registry-server" probeResult="failure" output=< Nov 28 14:59:12 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 14:59:12 crc kubenswrapper[4817]: > Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.870963 4817 scope.go:117] "RemoveContainer" containerID="06850586edffc8451a9e554b885a2782a2b8c9d6af6cc5b5ec8d9a4dfea1c71e" Nov 28 14:59:12 crc kubenswrapper[4817]: I1128 14:59:12.925231 4817 scope.go:117] "RemoveContainer" containerID="cf73eb05681a447f3f54901052a036fb35dfc99ed41bb400ebf920a3dca94a8f" Nov 28 14:59:13 crc kubenswrapper[4817]: I1128 14:59:13.022869 4817 scope.go:117] "RemoveContainer" containerID="9ebe3eab7e0802348099cb78a107daa6c1e074795a2cce2c2c9c80c8ed0bfc85" Nov 28 14:59:13 crc kubenswrapper[4817]: I1128 14:59:13.624334 4817 generic.go:334] "Generic (PLEG): container finished" podID="aa1fe5c5-0d23-47fb-990d-480cd18970b8" containerID="0473d4cd029e1ca139499dfa8b3eaaf6a1ea049651adb4c5aaf35cc9d1ad4f22" exitCode=0 Nov 28 14:59:13 crc kubenswrapper[4817]: I1128 14:59:13.624416 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" event={"ID":"aa1fe5c5-0d23-47fb-990d-480cd18970b8","Type":"ContainerDied","Data":"0473d4cd029e1ca139499dfa8b3eaaf6a1ea049651adb4c5aaf35cc9d1ad4f22"} Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.054407 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.076130 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-ssh-key\") pod \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.076191 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cdl6\" (UniqueName: \"kubernetes.io/projected/aa1fe5c5-0d23-47fb-990d-480cd18970b8-kube-api-access-9cdl6\") pod \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.076393 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-inventory\") pod \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\" (UID: \"aa1fe5c5-0d23-47fb-990d-480cd18970b8\") " Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.081669 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa1fe5c5-0d23-47fb-990d-480cd18970b8-kube-api-access-9cdl6" (OuterVolumeSpecName: "kube-api-access-9cdl6") pod "aa1fe5c5-0d23-47fb-990d-480cd18970b8" (UID: "aa1fe5c5-0d23-47fb-990d-480cd18970b8"). InnerVolumeSpecName "kube-api-access-9cdl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.105957 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-inventory" (OuterVolumeSpecName: "inventory") pod "aa1fe5c5-0d23-47fb-990d-480cd18970b8" (UID: "aa1fe5c5-0d23-47fb-990d-480cd18970b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.109849 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa1fe5c5-0d23-47fb-990d-480cd18970b8" (UID: "aa1fe5c5-0d23-47fb-990d-480cd18970b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.178861 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.178896 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa1fe5c5-0d23-47fb-990d-480cd18970b8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.178912 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cdl6\" (UniqueName: \"kubernetes.io/projected/aa1fe5c5-0d23-47fb-990d-480cd18970b8-kube-api-access-9cdl6\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.642835 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" event={"ID":"aa1fe5c5-0d23-47fb-990d-480cd18970b8","Type":"ContainerDied","Data":"35d803e81da7a524330d0d11f676b1413ca44b9f0e4c695f4623b8412867f56d"} Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.642872 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35d803e81da7a524330d0d11f676b1413ca44b9f0e4c695f4623b8412867f56d" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.642871 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.728377 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk"] Nov 28 14:59:15 crc kubenswrapper[4817]: E1128 14:59:15.728941 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa1fe5c5-0d23-47fb-990d-480cd18970b8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.728965 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa1fe5c5-0d23-47fb-990d-480cd18970b8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.729283 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa1fe5c5-0d23-47fb-990d-480cd18970b8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.730145 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.732239 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.732569 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.732782 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.733069 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.756712 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk"] Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.788085 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.788341 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9kzj\" (UniqueName: \"kubernetes.io/projected/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-kube-api-access-z9kzj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.788594 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.890834 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9kzj\" (UniqueName: \"kubernetes.io/projected/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-kube-api-access-z9kzj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.891392 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.891470 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.901215 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.903379 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:15 crc kubenswrapper[4817]: I1128 14:59:15.908115 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9kzj\" (UniqueName: \"kubernetes.io/projected/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-kube-api-access-z9kzj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-zd8vk\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:16 crc kubenswrapper[4817]: I1128 14:59:16.048466 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 14:59:16 crc kubenswrapper[4817]: I1128 14:59:16.610530 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk"] Nov 28 14:59:16 crc kubenswrapper[4817]: I1128 14:59:16.656121 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" event={"ID":"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91","Type":"ContainerStarted","Data":"ce1f140131c87dd22004a6d1db1340bf3943af277cf923f435ed43a60519302d"} Nov 28 14:59:20 crc kubenswrapper[4817]: I1128 14:59:20.235101 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" event={"ID":"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91","Type":"ContainerStarted","Data":"90294a50dc93532a0084d4d598b38b7a65066f9269e30ffb270f2a6277485c39"} Nov 28 14:59:21 crc kubenswrapper[4817]: I1128 14:59:21.264357 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" podStartSLOduration=5.025239798 podStartE2EDuration="6.26433811s" podCreationTimestamp="2025-11-28 14:59:15 +0000 UTC" firstStartedPulling="2025-11-28 14:59:16.623338674 +0000 UTC m=+1779.211316940" lastFinishedPulling="2025-11-28 14:59:17.862436976 +0000 UTC m=+1780.450415252" observedRunningTime="2025-11-28 14:59:21.255456898 +0000 UTC m=+1783.843435164" watchObservedRunningTime="2025-11-28 14:59:21.26433811 +0000 UTC m=+1783.852316376" Nov 28 14:59:21 crc kubenswrapper[4817]: I1128 14:59:21.737006 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:59:21 crc kubenswrapper[4817]: E1128 14:59:21.737298 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:59:21 crc kubenswrapper[4817]: I1128 14:59:21.764925 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:21 crc kubenswrapper[4817]: I1128 14:59:21.813085 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:22 crc kubenswrapper[4817]: I1128 14:59:22.003112 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzm8w"] Nov 28 14:59:22 crc kubenswrapper[4817]: I1128 14:59:22.041565 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-f52xr"] Nov 28 14:59:22 crc kubenswrapper[4817]: I1128 14:59:22.054143 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-f52xr"] Nov 28 14:59:23 crc kubenswrapper[4817]: I1128 14:59:23.264490 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jzm8w" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="registry-server" containerID="cri-o://149480f89d52150c634752a17b8cff67e6f8e7362e7e6bc915a941dbeac628d7" gracePeriod=2 Nov 28 14:59:23 crc kubenswrapper[4817]: I1128 14:59:23.748104 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c93d5fea-7f33-47a2-916a-80e9c68a89f2" path="/var/lib/kubelet/pods/c93d5fea-7f33-47a2-916a-80e9c68a89f2/volumes" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.291711 4817 generic.go:334] "Generic (PLEG): container finished" podID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerID="149480f89d52150c634752a17b8cff67e6f8e7362e7e6bc915a941dbeac628d7" exitCode=0 Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.291810 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerDied","Data":"149480f89d52150c634752a17b8cff67e6f8e7362e7e6bc915a941dbeac628d7"} Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.470346 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.496366 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-utilities\") pod \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.496492 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8xrw\" (UniqueName: \"kubernetes.io/projected/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-kube-api-access-q8xrw\") pod \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.496775 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-catalog-content\") pod \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\" (UID: \"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa\") " Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.497968 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-utilities" (OuterVolumeSpecName: "utilities") pod "31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" (UID: "31fff6f2-9a2d-4b2f-b106-9043c20ac3fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.503535 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-kube-api-access-q8xrw" (OuterVolumeSpecName: "kube-api-access-q8xrw") pod "31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" (UID: "31fff6f2-9a2d-4b2f-b106-9043c20ac3fa"). InnerVolumeSpecName "kube-api-access-q8xrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.599211 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.599248 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8xrw\" (UniqueName: \"kubernetes.io/projected/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-kube-api-access-q8xrw\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.629104 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" (UID: "31fff6f2-9a2d-4b2f-b106-9043c20ac3fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:59:24 crc kubenswrapper[4817]: I1128 14:59:24.701575 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.309533 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzm8w" event={"ID":"31fff6f2-9a2d-4b2f-b106-9043c20ac3fa","Type":"ContainerDied","Data":"9f57f1714f0895169fa02d6f28b6ecacd8ad6972986e1e6ff98d753350cba77a"} Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.309923 4817 scope.go:117] "RemoveContainer" containerID="149480f89d52150c634752a17b8cff67e6f8e7362e7e6bc915a941dbeac628d7" Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.309625 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzm8w" Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.353646 4817 scope.go:117] "RemoveContainer" containerID="c1927cff2f7c8deaf1e952bacc31a6abf309bec781ad33059e72ca6edd8b6a09" Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.365897 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzm8w"] Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.374698 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jzm8w"] Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.403284 4817 scope.go:117] "RemoveContainer" containerID="6f9242d083e6cb9425de5393c80785e1822c505bd7489e193c587b902ca3a50d" Nov 28 14:59:25 crc kubenswrapper[4817]: I1128 14:59:25.752092 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" path="/var/lib/kubelet/pods/31fff6f2-9a2d-4b2f-b106-9043c20ac3fa/volumes" Nov 28 14:59:26 crc kubenswrapper[4817]: I1128 14:59:26.052368 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q994c"] Nov 28 14:59:26 crc kubenswrapper[4817]: I1128 14:59:26.071550 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q994c"] Nov 28 14:59:27 crc kubenswrapper[4817]: I1128 14:59:27.786530 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2f8c40-8b05-4037-8add-58252a9f996d" path="/var/lib/kubelet/pods/1b2f8c40-8b05-4037-8add-58252a9f996d/volumes" Nov 28 14:59:34 crc kubenswrapper[4817]: I1128 14:59:34.738211 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:59:34 crc kubenswrapper[4817]: E1128 14:59:34.739040 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:59:47 crc kubenswrapper[4817]: I1128 14:59:47.743060 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:59:47 crc kubenswrapper[4817]: E1128 14:59:47.744064 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.080479 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4nqjs"] Nov 28 14:59:52 crc kubenswrapper[4817]: E1128 14:59:52.081410 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="extract-utilities" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.081428 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="extract-utilities" Nov 28 14:59:52 crc kubenswrapper[4817]: E1128 14:59:52.081455 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="extract-content" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.081464 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="extract-content" Nov 28 14:59:52 crc kubenswrapper[4817]: E1128 14:59:52.081487 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="registry-server" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.081496 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="registry-server" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.081751 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="31fff6f2-9a2d-4b2f-b106-9043c20ac3fa" containerName="registry-server" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.084137 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.090014 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4nqjs"] Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.223875 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-utilities\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.223935 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n64mp\" (UniqueName: \"kubernetes.io/projected/cb9e18a8-c259-4d68-a11a-73c64a578262-kube-api-access-n64mp\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.223963 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-catalog-content\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.325642 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-utilities\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.325696 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n64mp\" (UniqueName: \"kubernetes.io/projected/cb9e18a8-c259-4d68-a11a-73c64a578262-kube-api-access-n64mp\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.325734 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-catalog-content\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.326179 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-utilities\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.326359 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-catalog-content\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.354867 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n64mp\" (UniqueName: \"kubernetes.io/projected/cb9e18a8-c259-4d68-a11a-73c64a578262-kube-api-access-n64mp\") pod \"community-operators-4nqjs\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.415033 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 14:59:52 crc kubenswrapper[4817]: I1128 14:59:52.954741 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4nqjs"] Nov 28 14:59:53 crc kubenswrapper[4817]: I1128 14:59:53.665869 4817 generic.go:334] "Generic (PLEG): container finished" podID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerID="0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854" exitCode=0 Nov 28 14:59:53 crc kubenswrapper[4817]: I1128 14:59:53.665915 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerDied","Data":"0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854"} Nov 28 14:59:53 crc kubenswrapper[4817]: I1128 14:59:53.665954 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerStarted","Data":"a85783b425b7dbfd5ddab5e9f6353e13a90827b9f082c844d3cb70f3d9b6a166"} Nov 28 14:59:54 crc kubenswrapper[4817]: I1128 14:59:54.678388 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerStarted","Data":"c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790"} Nov 28 14:59:55 crc kubenswrapper[4817]: I1128 14:59:55.696423 4817 generic.go:334] "Generic (PLEG): container finished" podID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerID="c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790" exitCode=0 Nov 28 14:59:55 crc kubenswrapper[4817]: I1128 14:59:55.696507 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerDied","Data":"c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790"} Nov 28 14:59:58 crc kubenswrapper[4817]: I1128 14:59:58.730764 4817 generic.go:334] "Generic (PLEG): container finished" podID="62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" containerID="90294a50dc93532a0084d4d598b38b7a65066f9269e30ffb270f2a6277485c39" exitCode=0 Nov 28 14:59:58 crc kubenswrapper[4817]: I1128 14:59:58.730886 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" event={"ID":"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91","Type":"ContainerDied","Data":"90294a50dc93532a0084d4d598b38b7a65066f9269e30ffb270f2a6277485c39"} Nov 28 14:59:59 crc kubenswrapper[4817]: I1128 14:59:59.738432 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 14:59:59 crc kubenswrapper[4817]: E1128 14:59:59.739480 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 14:59:59 crc kubenswrapper[4817]: I1128 14:59:59.772059 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerStarted","Data":"82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f"} Nov 28 14:59:59 crc kubenswrapper[4817]: I1128 14:59:59.782361 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4nqjs" podStartSLOduration=2.708913625 podStartE2EDuration="7.782334702s" podCreationTimestamp="2025-11-28 14:59:52 +0000 UTC" firstStartedPulling="2025-11-28 14:59:53.671495051 +0000 UTC m=+1816.259473317" lastFinishedPulling="2025-11-28 14:59:58.744916118 +0000 UTC m=+1821.332894394" observedRunningTime="2025-11-28 14:59:59.775551953 +0000 UTC m=+1822.363530229" watchObservedRunningTime="2025-11-28 14:59:59.782334702 +0000 UTC m=+1822.370312988" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.155238 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp"] Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.157008 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.158877 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.159955 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.210232 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp"] Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.244294 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.298617 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64wf2\" (UniqueName: \"kubernetes.io/projected/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-kube-api-access-64wf2\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.298698 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-secret-volume\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.299095 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-config-volume\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.400636 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-ssh-key\") pod \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.400843 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-inventory\") pod \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.401022 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9kzj\" (UniqueName: \"kubernetes.io/projected/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-kube-api-access-z9kzj\") pod \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\" (UID: \"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91\") " Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.401478 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-config-volume\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.401545 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64wf2\" (UniqueName: \"kubernetes.io/projected/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-kube-api-access-64wf2\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.401633 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-secret-volume\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.403091 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-config-volume\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.406789 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-kube-api-access-z9kzj" (OuterVolumeSpecName: "kube-api-access-z9kzj") pod "62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" (UID: "62cd0765-6d4e-4e36-96df-8e3cc4bf8a91"). InnerVolumeSpecName "kube-api-access-z9kzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.421057 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-secret-volume\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.421807 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64wf2\" (UniqueName: \"kubernetes.io/projected/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-kube-api-access-64wf2\") pod \"collect-profiles-29405700-gzgbp\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.436427 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" (UID: "62cd0765-6d4e-4e36-96df-8e3cc4bf8a91"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.437029 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-inventory" (OuterVolumeSpecName: "inventory") pod "62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" (UID: "62cd0765-6d4e-4e36-96df-8e3cc4bf8a91"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.503850 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9kzj\" (UniqueName: \"kubernetes.io/projected/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-kube-api-access-z9kzj\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.503885 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.503897 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62cd0765-6d4e-4e36-96df-8e3cc4bf8a91-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.557226 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.759545 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.765214 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-zd8vk" event={"ID":"62cd0765-6d4e-4e36-96df-8e3cc4bf8a91","Type":"ContainerDied","Data":"ce1f140131c87dd22004a6d1db1340bf3943af277cf923f435ed43a60519302d"} Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.765251 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce1f140131c87dd22004a6d1db1340bf3943af277cf923f435ed43a60519302d" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.840795 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8"] Nov 28 15:00:00 crc kubenswrapper[4817]: E1128 15:00:00.841522 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.841541 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.841762 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="62cd0765-6d4e-4e36-96df-8e3cc4bf8a91" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.842481 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.845829 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.846047 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.846324 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.846492 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.858557 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8"] Nov 28 15:00:00 crc kubenswrapper[4817]: W1128 15:00:00.985019 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10b209ee_9bd3_4e8b_a1a1_f74e08a431f6.slice/crio-7745d6ad803fd25d4ff61a73fea6c7c5d819b0ea18fc8c5fd02bda29d129c164 WatchSource:0}: Error finding container 7745d6ad803fd25d4ff61a73fea6c7c5d819b0ea18fc8c5fd02bda29d129c164: Status 404 returned error can't find the container with id 7745d6ad803fd25d4ff61a73fea6c7c5d819b0ea18fc8c5fd02bda29d129c164 Nov 28 15:00:00 crc kubenswrapper[4817]: I1128 15:00:00.991994 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp"] Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.014424 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.014875 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.015169 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbzvp\" (UniqueName: \"kubernetes.io/projected/1a488bc2-89d2-49a6-88a3-0656eb5deb56-kube-api-access-cbzvp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.117222 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.117370 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.117405 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbzvp\" (UniqueName: \"kubernetes.io/projected/1a488bc2-89d2-49a6-88a3-0656eb5deb56-kube-api-access-cbzvp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.125808 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.127349 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.136214 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbzvp\" (UniqueName: \"kubernetes.io/projected/1a488bc2-89d2-49a6-88a3-0656eb5deb56-kube-api-access-cbzvp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.162400 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.772708 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" event={"ID":"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6","Type":"ContainerStarted","Data":"120f5c697586aeaa59bcd06caeee7c552acae7a9ab3422fb9a50a7afb89a0477"} Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.773504 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" event={"ID":"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6","Type":"ContainerStarted","Data":"7745d6ad803fd25d4ff61a73fea6c7c5d819b0ea18fc8c5fd02bda29d129c164"} Nov 28 15:00:01 crc kubenswrapper[4817]: I1128 15:00:01.813076 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" podStartSLOduration=1.8130533 podStartE2EDuration="1.8130533s" podCreationTimestamp="2025-11-28 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:00:01.793161234 +0000 UTC m=+1824.381139500" watchObservedRunningTime="2025-11-28 15:00:01.8130533 +0000 UTC m=+1824.401031566" Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.116649 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8"] Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.415608 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.416000 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.472488 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.786528 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" event={"ID":"1a488bc2-89d2-49a6-88a3-0656eb5deb56","Type":"ContainerStarted","Data":"f15e8864b3eaf76efa62bf1c84afdb17998511dfa2f8c8b57485d486eefb82a5"} Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.789038 4817 generic.go:334] "Generic (PLEG): container finished" podID="10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" containerID="120f5c697586aeaa59bcd06caeee7c552acae7a9ab3422fb9a50a7afb89a0477" exitCode=0 Nov 28 15:00:02 crc kubenswrapper[4817]: I1128 15:00:02.789587 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" event={"ID":"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6","Type":"ContainerDied","Data":"120f5c697586aeaa59bcd06caeee7c552acae7a9ab3422fb9a50a7afb89a0477"} Nov 28 15:00:03 crc kubenswrapper[4817]: I1128 15:00:03.803159 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" event={"ID":"1a488bc2-89d2-49a6-88a3-0656eb5deb56","Type":"ContainerStarted","Data":"824a4d03ef11d50602e939fc2f474dce7e499ab5a0321a3e7b731102b8d9f17d"} Nov 28 15:00:03 crc kubenswrapper[4817]: I1128 15:00:03.831043 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" podStartSLOduration=3.155348246 podStartE2EDuration="3.831013291s" podCreationTimestamp="2025-11-28 15:00:00 +0000 UTC" firstStartedPulling="2025-11-28 15:00:02.127029188 +0000 UTC m=+1824.715007454" lastFinishedPulling="2025-11-28 15:00:02.802694223 +0000 UTC m=+1825.390672499" observedRunningTime="2025-11-28 15:00:03.825322329 +0000 UTC m=+1826.413300635" watchObservedRunningTime="2025-11-28 15:00:03.831013291 +0000 UTC m=+1826.418991607" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.191503 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.283386 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-config-volume\") pod \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.283524 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-secret-volume\") pod \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.283655 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64wf2\" (UniqueName: \"kubernetes.io/projected/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-kube-api-access-64wf2\") pod \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\" (UID: \"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6\") " Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.284319 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-config-volume" (OuterVolumeSpecName: "config-volume") pod "10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" (UID: "10b209ee-9bd3-4e8b-a1a1-f74e08a431f6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.291443 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-kube-api-access-64wf2" (OuterVolumeSpecName: "kube-api-access-64wf2") pod "10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" (UID: "10b209ee-9bd3-4e8b-a1a1-f74e08a431f6"). InnerVolumeSpecName "kube-api-access-64wf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.292573 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" (UID: "10b209ee-9bd3-4e8b-a1a1-f74e08a431f6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.386119 4817 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.386147 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64wf2\" (UniqueName: \"kubernetes.io/projected/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-kube-api-access-64wf2\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.386157 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.819225 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.819218 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp" event={"ID":"10b209ee-9bd3-4e8b-a1a1-f74e08a431f6","Type":"ContainerDied","Data":"7745d6ad803fd25d4ff61a73fea6c7c5d819b0ea18fc8c5fd02bda29d129c164"} Nov 28 15:00:04 crc kubenswrapper[4817]: I1128 15:00:04.819348 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7745d6ad803fd25d4ff61a73fea6c7c5d819b0ea18fc8c5fd02bda29d129c164" Nov 28 15:00:06 crc kubenswrapper[4817]: I1128 15:00:06.051821 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dfjsl"] Nov 28 15:00:06 crc kubenswrapper[4817]: I1128 15:00:06.060022 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dfjsl"] Nov 28 15:00:07 crc kubenswrapper[4817]: I1128 15:00:07.758320 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d23a9c3-63c8-49a7-80e2-7f6148e0bf06" path="/var/lib/kubelet/pods/0d23a9c3-63c8-49a7-80e2-7f6148e0bf06/volumes" Nov 28 15:00:12 crc kubenswrapper[4817]: I1128 15:00:12.471523 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 15:00:12 crc kubenswrapper[4817]: I1128 15:00:12.523361 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4nqjs"] Nov 28 15:00:12 crc kubenswrapper[4817]: I1128 15:00:12.737223 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:00:12 crc kubenswrapper[4817]: E1128 15:00:12.737673 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:00:12 crc kubenswrapper[4817]: I1128 15:00:12.910862 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4nqjs" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="registry-server" containerID="cri-o://82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f" gracePeriod=2 Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.148059 4817 scope.go:117] "RemoveContainer" containerID="a4d4ba8ce44f93b94b60753c95cc2c62a36748490a2bbf51e4bb2d6e7abbd5c2" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.203363 4817 scope.go:117] "RemoveContainer" containerID="18e8370affb31e5bb70b35b5de7775b06051710af0ef2e5cc37bf860c16698f6" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.273825 4817 scope.go:117] "RemoveContainer" containerID="53266733de55237f9fe7dbd93ec510da9273166899b424c157266d18d42e5df4" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.379046 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.487781 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-catalog-content\") pod \"cb9e18a8-c259-4d68-a11a-73c64a578262\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.487866 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n64mp\" (UniqueName: \"kubernetes.io/projected/cb9e18a8-c259-4d68-a11a-73c64a578262-kube-api-access-n64mp\") pod \"cb9e18a8-c259-4d68-a11a-73c64a578262\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.487896 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-utilities\") pod \"cb9e18a8-c259-4d68-a11a-73c64a578262\" (UID: \"cb9e18a8-c259-4d68-a11a-73c64a578262\") " Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.488711 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-utilities" (OuterVolumeSpecName: "utilities") pod "cb9e18a8-c259-4d68-a11a-73c64a578262" (UID: "cb9e18a8-c259-4d68-a11a-73c64a578262"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.494111 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb9e18a8-c259-4d68-a11a-73c64a578262-kube-api-access-n64mp" (OuterVolumeSpecName: "kube-api-access-n64mp") pod "cb9e18a8-c259-4d68-a11a-73c64a578262" (UID: "cb9e18a8-c259-4d68-a11a-73c64a578262"). InnerVolumeSpecName "kube-api-access-n64mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.537902 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb9e18a8-c259-4d68-a11a-73c64a578262" (UID: "cb9e18a8-c259-4d68-a11a-73c64a578262"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.589432 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.589465 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n64mp\" (UniqueName: \"kubernetes.io/projected/cb9e18a8-c259-4d68-a11a-73c64a578262-kube-api-access-n64mp\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.589478 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9e18a8-c259-4d68-a11a-73c64a578262-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.936818 4817 generic.go:334] "Generic (PLEG): container finished" podID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerID="82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f" exitCode=0 Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.937017 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nqjs" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.937025 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerDied","Data":"82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f"} Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.939810 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nqjs" event={"ID":"cb9e18a8-c259-4d68-a11a-73c64a578262","Type":"ContainerDied","Data":"a85783b425b7dbfd5ddab5e9f6353e13a90827b9f082c844d3cb70f3d9b6a166"} Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.939999 4817 scope.go:117] "RemoveContainer" containerID="82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.981221 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4nqjs"] Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.986624 4817 scope.go:117] "RemoveContainer" containerID="c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790" Nov 28 15:00:13 crc kubenswrapper[4817]: I1128 15:00:13.988707 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4nqjs"] Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.022307 4817 scope.go:117] "RemoveContainer" containerID="0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854" Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.059740 4817 scope.go:117] "RemoveContainer" containerID="82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f" Nov 28 15:00:14 crc kubenswrapper[4817]: E1128 15:00:14.060974 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f\": container with ID starting with 82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f not found: ID does not exist" containerID="82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f" Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.061010 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f"} err="failed to get container status \"82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f\": rpc error: code = NotFound desc = could not find container \"82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f\": container with ID starting with 82a6e44c72a3ad3b8356b93558c67b0ca40d4499661646f43a37d1508f06746f not found: ID does not exist" Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.061033 4817 scope.go:117] "RemoveContainer" containerID="c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790" Nov 28 15:00:14 crc kubenswrapper[4817]: E1128 15:00:14.061404 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790\": container with ID starting with c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790 not found: ID does not exist" containerID="c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790" Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.061428 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790"} err="failed to get container status \"c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790\": rpc error: code = NotFound desc = could not find container \"c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790\": container with ID starting with c8e829ccf990e2ca6bf2ded81b0bd886d220c32f19efea0560b66e0bf1ae0790 not found: ID does not exist" Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.061446 4817 scope.go:117] "RemoveContainer" containerID="0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854" Nov 28 15:00:14 crc kubenswrapper[4817]: E1128 15:00:14.061635 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854\": container with ID starting with 0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854 not found: ID does not exist" containerID="0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854" Nov 28 15:00:14 crc kubenswrapper[4817]: I1128 15:00:14.061657 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854"} err="failed to get container status \"0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854\": rpc error: code = NotFound desc = could not find container \"0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854\": container with ID starting with 0a622b2afb9ff2b56d79b77f36fa1feeb81f56b8ce66c36f0d7c8a7763426854 not found: ID does not exist" Nov 28 15:00:15 crc kubenswrapper[4817]: I1128 15:00:15.747898 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" path="/var/lib/kubelet/pods/cb9e18a8-c259-4d68-a11a-73c64a578262/volumes" Nov 28 15:00:23 crc kubenswrapper[4817]: I1128 15:00:23.737327 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:00:23 crc kubenswrapper[4817]: E1128 15:00:23.737931 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:00:35 crc kubenswrapper[4817]: I1128 15:00:35.737293 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:00:35 crc kubenswrapper[4817]: E1128 15:00:35.738059 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:00:48 crc kubenswrapper[4817]: I1128 15:00:48.737377 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:00:48 crc kubenswrapper[4817]: E1128 15:00:48.738508 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:00:57 crc kubenswrapper[4817]: I1128 15:00:57.392681 4817 generic.go:334] "Generic (PLEG): container finished" podID="1a488bc2-89d2-49a6-88a3-0656eb5deb56" containerID="824a4d03ef11d50602e939fc2f474dce7e499ab5a0321a3e7b731102b8d9f17d" exitCode=0 Nov 28 15:00:57 crc kubenswrapper[4817]: I1128 15:00:57.392780 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" event={"ID":"1a488bc2-89d2-49a6-88a3-0656eb5deb56","Type":"ContainerDied","Data":"824a4d03ef11d50602e939fc2f474dce7e499ab5a0321a3e7b731102b8d9f17d"} Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.808074 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.917266 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-inventory\") pod \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.917320 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbzvp\" (UniqueName: \"kubernetes.io/projected/1a488bc2-89d2-49a6-88a3-0656eb5deb56-kube-api-access-cbzvp\") pod \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.917561 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-ssh-key\") pod \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\" (UID: \"1a488bc2-89d2-49a6-88a3-0656eb5deb56\") " Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.922598 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a488bc2-89d2-49a6-88a3-0656eb5deb56-kube-api-access-cbzvp" (OuterVolumeSpecName: "kube-api-access-cbzvp") pod "1a488bc2-89d2-49a6-88a3-0656eb5deb56" (UID: "1a488bc2-89d2-49a6-88a3-0656eb5deb56"). InnerVolumeSpecName "kube-api-access-cbzvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.943455 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a488bc2-89d2-49a6-88a3-0656eb5deb56" (UID: "1a488bc2-89d2-49a6-88a3-0656eb5deb56"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:00:58 crc kubenswrapper[4817]: I1128 15:00:58.958445 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-inventory" (OuterVolumeSpecName: "inventory") pod "1a488bc2-89d2-49a6-88a3-0656eb5deb56" (UID: "1a488bc2-89d2-49a6-88a3-0656eb5deb56"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.019615 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.019651 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a488bc2-89d2-49a6-88a3-0656eb5deb56-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.019660 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbzvp\" (UniqueName: \"kubernetes.io/projected/1a488bc2-89d2-49a6-88a3-0656eb5deb56-kube-api-access-cbzvp\") on node \"crc\" DevicePath \"\"" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.412076 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" event={"ID":"1a488bc2-89d2-49a6-88a3-0656eb5deb56","Type":"ContainerDied","Data":"f15e8864b3eaf76efa62bf1c84afdb17998511dfa2f8c8b57485d486eefb82a5"} Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.412106 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.412120 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f15e8864b3eaf76efa62bf1c84afdb17998511dfa2f8c8b57485d486eefb82a5" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.502766 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jzgbr"] Nov 28 15:00:59 crc kubenswrapper[4817]: E1128 15:00:59.503163 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a488bc2-89d2-49a6-88a3-0656eb5deb56" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503189 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a488bc2-89d2-49a6-88a3-0656eb5deb56" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:00:59 crc kubenswrapper[4817]: E1128 15:00:59.503206 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="registry-server" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503213 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="registry-server" Nov 28 15:00:59 crc kubenswrapper[4817]: E1128 15:00:59.503238 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="extract-content" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503246 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="extract-content" Nov 28 15:00:59 crc kubenswrapper[4817]: E1128 15:00:59.503258 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="extract-utilities" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503266 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="extract-utilities" Nov 28 15:00:59 crc kubenswrapper[4817]: E1128 15:00:59.503284 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" containerName="collect-profiles" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503293 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" containerName="collect-profiles" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503491 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" containerName="collect-profiles" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503520 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a488bc2-89d2-49a6-88a3-0656eb5deb56" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.503535 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9e18a8-c259-4d68-a11a-73c64a578262" containerName="registry-server" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.504401 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.506740 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.506804 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.506848 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.510371 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.514854 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jzgbr"] Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.627490 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.627579 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqzl7\" (UniqueName: \"kubernetes.io/projected/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-kube-api-access-bqzl7\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.627620 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.729026 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqzl7\" (UniqueName: \"kubernetes.io/projected/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-kube-api-access-bqzl7\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.729093 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.729185 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.734540 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.737902 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:00:59 crc kubenswrapper[4817]: E1128 15:00:59.738303 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.739199 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.755665 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqzl7\" (UniqueName: \"kubernetes.io/projected/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-kube-api-access-bqzl7\") pod \"ssh-known-hosts-edpm-deployment-jzgbr\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:00:59 crc kubenswrapper[4817]: I1128 15:00:59.821428 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.138875 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29405701-7q48x"] Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.140553 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.158366 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29405701-7q48x"] Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.163410 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-config-data\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.163498 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tdvx\" (UniqueName: \"kubernetes.io/projected/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-kube-api-access-8tdvx\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.163529 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-fernet-keys\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.163562 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-combined-ca-bundle\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.264895 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-config-data\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.264993 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tdvx\" (UniqueName: \"kubernetes.io/projected/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-kube-api-access-8tdvx\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.265033 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-fernet-keys\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.265085 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-combined-ca-bundle\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.277660 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-combined-ca-bundle\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.282699 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-fernet-keys\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.296549 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-config-data\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.297493 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tdvx\" (UniqueName: \"kubernetes.io/projected/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-kube-api-access-8tdvx\") pod \"keystone-cron-29405701-7q48x\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.463402 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jzgbr"] Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.467497 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.468271 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:00 crc kubenswrapper[4817]: I1128 15:01:00.882909 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29405701-7q48x"] Nov 28 15:01:00 crc kubenswrapper[4817]: W1128 15:01:00.890758 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ef2895d_d3df_4a6b_abad_c7d1bceb0871.slice/crio-7caac18ea2c1980bf6b573da707e6f4e63a35b7cc173672676694609011cb28b WatchSource:0}: Error finding container 7caac18ea2c1980bf6b573da707e6f4e63a35b7cc173672676694609011cb28b: Status 404 returned error can't find the container with id 7caac18ea2c1980bf6b573da707e6f4e63a35b7cc173672676694609011cb28b Nov 28 15:01:01 crc kubenswrapper[4817]: I1128 15:01:01.434913 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405701-7q48x" event={"ID":"9ef2895d-d3df-4a6b-abad-c7d1bceb0871","Type":"ContainerStarted","Data":"b22ae7942573c141f7ea994d867ddde19173a114de13e56e855557e4773b794f"} Nov 28 15:01:01 crc kubenswrapper[4817]: I1128 15:01:01.435303 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405701-7q48x" event={"ID":"9ef2895d-d3df-4a6b-abad-c7d1bceb0871","Type":"ContainerStarted","Data":"7caac18ea2c1980bf6b573da707e6f4e63a35b7cc173672676694609011cb28b"} Nov 28 15:01:01 crc kubenswrapper[4817]: I1128 15:01:01.437705 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" event={"ID":"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad","Type":"ContainerStarted","Data":"3ba7106af5604841a72c1106fd0cda1c17b4f344526f4469d9e2f888e7b793a8"} Nov 28 15:01:01 crc kubenswrapper[4817]: I1128 15:01:01.462269 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29405701-7q48x" podStartSLOduration=1.4621709090000001 podStartE2EDuration="1.462170909s" podCreationTimestamp="2025-11-28 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:01:01.453684107 +0000 UTC m=+1884.041662373" watchObservedRunningTime="2025-11-28 15:01:01.462170909 +0000 UTC m=+1884.050149175" Nov 28 15:01:02 crc kubenswrapper[4817]: I1128 15:01:02.454405 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" event={"ID":"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad","Type":"ContainerStarted","Data":"7b86acedeec9e9c163e737328cf141ec4759d0bdc4756d04395eb29f49ce4a1d"} Nov 28 15:01:02 crc kubenswrapper[4817]: I1128 15:01:02.479391 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" podStartSLOduration=2.596682483 podStartE2EDuration="3.479354069s" podCreationTimestamp="2025-11-28 15:00:59 +0000 UTC" firstStartedPulling="2025-11-28 15:01:00.467200213 +0000 UTC m=+1883.055178479" lastFinishedPulling="2025-11-28 15:01:01.349871799 +0000 UTC m=+1883.937850065" observedRunningTime="2025-11-28 15:01:02.475164064 +0000 UTC m=+1885.063142370" watchObservedRunningTime="2025-11-28 15:01:02.479354069 +0000 UTC m=+1885.067332375" Nov 28 15:01:03 crc kubenswrapper[4817]: I1128 15:01:03.462847 4817 generic.go:334] "Generic (PLEG): container finished" podID="9ef2895d-d3df-4a6b-abad-c7d1bceb0871" containerID="b22ae7942573c141f7ea994d867ddde19173a114de13e56e855557e4773b794f" exitCode=0 Nov 28 15:01:03 crc kubenswrapper[4817]: I1128 15:01:03.462925 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405701-7q48x" event={"ID":"9ef2895d-d3df-4a6b-abad-c7d1bceb0871","Type":"ContainerDied","Data":"b22ae7942573c141f7ea994d867ddde19173a114de13e56e855557e4773b794f"} Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.788530 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.874040 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdvx\" (UniqueName: \"kubernetes.io/projected/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-kube-api-access-8tdvx\") pod \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.874128 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-combined-ca-bundle\") pod \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.874152 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-config-data\") pod \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.874168 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-fernet-keys\") pod \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\" (UID: \"9ef2895d-d3df-4a6b-abad-c7d1bceb0871\") " Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.878940 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-kube-api-access-8tdvx" (OuterVolumeSpecName: "kube-api-access-8tdvx") pod "9ef2895d-d3df-4a6b-abad-c7d1bceb0871" (UID: "9ef2895d-d3df-4a6b-abad-c7d1bceb0871"). InnerVolumeSpecName "kube-api-access-8tdvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.884181 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9ef2895d-d3df-4a6b-abad-c7d1bceb0871" (UID: "9ef2895d-d3df-4a6b-abad-c7d1bceb0871"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.902686 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ef2895d-d3df-4a6b-abad-c7d1bceb0871" (UID: "9ef2895d-d3df-4a6b-abad-c7d1bceb0871"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.925574 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-config-data" (OuterVolumeSpecName: "config-data") pod "9ef2895d-d3df-4a6b-abad-c7d1bceb0871" (UID: "9ef2895d-d3df-4a6b-abad-c7d1bceb0871"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.975882 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdvx\" (UniqueName: \"kubernetes.io/projected/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-kube-api-access-8tdvx\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.975914 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.975924 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:04 crc kubenswrapper[4817]: I1128 15:01:04.975933 4817 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9ef2895d-d3df-4a6b-abad-c7d1bceb0871-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:05 crc kubenswrapper[4817]: I1128 15:01:05.483245 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405701-7q48x" event={"ID":"9ef2895d-d3df-4a6b-abad-c7d1bceb0871","Type":"ContainerDied","Data":"7caac18ea2c1980bf6b573da707e6f4e63a35b7cc173672676694609011cb28b"} Nov 28 15:01:05 crc kubenswrapper[4817]: I1128 15:01:05.483287 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7caac18ea2c1980bf6b573da707e6f4e63a35b7cc173672676694609011cb28b" Nov 28 15:01:05 crc kubenswrapper[4817]: I1128 15:01:05.483367 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405701-7q48x" Nov 28 15:01:08 crc kubenswrapper[4817]: I1128 15:01:08.518519 4817 generic.go:334] "Generic (PLEG): container finished" podID="3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" containerID="7b86acedeec9e9c163e737328cf141ec4759d0bdc4756d04395eb29f49ce4a1d" exitCode=0 Nov 28 15:01:08 crc kubenswrapper[4817]: I1128 15:01:08.518613 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" event={"ID":"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad","Type":"ContainerDied","Data":"7b86acedeec9e9c163e737328cf141ec4759d0bdc4756d04395eb29f49ce4a1d"} Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.004034 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.069685 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-inventory-0\") pod \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.069767 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-ssh-key-openstack-edpm-ipam\") pod \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.069964 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqzl7\" (UniqueName: \"kubernetes.io/projected/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-kube-api-access-bqzl7\") pod \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\" (UID: \"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad\") " Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.075229 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-kube-api-access-bqzl7" (OuterVolumeSpecName: "kube-api-access-bqzl7") pod "3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" (UID: "3867da3b-518a-4aa7-80b4-5d6e5f4e7dad"). InnerVolumeSpecName "kube-api-access-bqzl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.100008 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" (UID: "3867da3b-518a-4aa7-80b4-5d6e5f4e7dad"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.108063 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" (UID: "3867da3b-518a-4aa7-80b4-5d6e5f4e7dad"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.172075 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqzl7\" (UniqueName: \"kubernetes.io/projected/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-kube-api-access-bqzl7\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.172109 4817 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.172119 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3867da3b-518a-4aa7-80b4-5d6e5f4e7dad-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.541190 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" event={"ID":"3867da3b-518a-4aa7-80b4-5d6e5f4e7dad","Type":"ContainerDied","Data":"3ba7106af5604841a72c1106fd0cda1c17b4f344526f4469d9e2f888e7b793a8"} Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.541229 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ba7106af5604841a72c1106fd0cda1c17b4f344526f4469d9e2f888e7b793a8" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.541273 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jzgbr" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.635663 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld"] Nov 28 15:01:10 crc kubenswrapper[4817]: E1128 15:01:10.636087 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef2895d-d3df-4a6b-abad-c7d1bceb0871" containerName="keystone-cron" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.636105 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef2895d-d3df-4a6b-abad-c7d1bceb0871" containerName="keystone-cron" Nov 28 15:01:10 crc kubenswrapper[4817]: E1128 15:01:10.636137 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" containerName="ssh-known-hosts-edpm-deployment" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.636144 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" containerName="ssh-known-hosts-edpm-deployment" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.636362 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3867da3b-518a-4aa7-80b4-5d6e5f4e7dad" containerName="ssh-known-hosts-edpm-deployment" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.636378 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef2895d-d3df-4a6b-abad-c7d1bceb0871" containerName="keystone-cron" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.637204 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.640541 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.641672 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.641986 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.642217 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.659586 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld"] Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.683564 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kmgb\" (UniqueName: \"kubernetes.io/projected/0b6cd940-c79b-4bd7-80ff-69bb850709ae-kube-api-access-7kmgb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.683670 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.683817 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.785911 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.785969 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kmgb\" (UniqueName: \"kubernetes.io/projected/0b6cd940-c79b-4bd7-80ff-69bb850709ae-kube-api-access-7kmgb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.786065 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.797778 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.798288 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.817412 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kmgb\" (UniqueName: \"kubernetes.io/projected/0b6cd940-c79b-4bd7-80ff-69bb850709ae-kube-api-access-7kmgb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvgld\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:10 crc kubenswrapper[4817]: I1128 15:01:10.973212 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:11 crc kubenswrapper[4817]: I1128 15:01:11.369227 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld"] Nov 28 15:01:11 crc kubenswrapper[4817]: I1128 15:01:11.550629 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" event={"ID":"0b6cd940-c79b-4bd7-80ff-69bb850709ae","Type":"ContainerStarted","Data":"f93da301de2aa28a68b294223f81cb4f2d78c5a9c60b7a28235c9d74305d0abc"} Nov 28 15:01:12 crc kubenswrapper[4817]: I1128 15:01:12.564626 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" event={"ID":"0b6cd940-c79b-4bd7-80ff-69bb850709ae","Type":"ContainerStarted","Data":"7a9b19c5601b850a6065396ca9809789afe2ea569e65c5a42e48bb26c5bb8d84"} Nov 28 15:01:12 crc kubenswrapper[4817]: I1128 15:01:12.597676 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" podStartSLOduration=2.048681714 podStartE2EDuration="2.597654271s" podCreationTimestamp="2025-11-28 15:01:10 +0000 UTC" firstStartedPulling="2025-11-28 15:01:11.38309365 +0000 UTC m=+1893.971071916" lastFinishedPulling="2025-11-28 15:01:11.932066207 +0000 UTC m=+1894.520044473" observedRunningTime="2025-11-28 15:01:12.592682767 +0000 UTC m=+1895.180661033" watchObservedRunningTime="2025-11-28 15:01:12.597654271 +0000 UTC m=+1895.185632557" Nov 28 15:01:13 crc kubenswrapper[4817]: I1128 15:01:13.737162 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:01:13 crc kubenswrapper[4817]: E1128 15:01:13.737800 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:01:20 crc kubenswrapper[4817]: I1128 15:01:20.666101 4817 generic.go:334] "Generic (PLEG): container finished" podID="0b6cd940-c79b-4bd7-80ff-69bb850709ae" containerID="7a9b19c5601b850a6065396ca9809789afe2ea569e65c5a42e48bb26c5bb8d84" exitCode=0 Nov 28 15:01:20 crc kubenswrapper[4817]: I1128 15:01:20.666193 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" event={"ID":"0b6cd940-c79b-4bd7-80ff-69bb850709ae","Type":"ContainerDied","Data":"7a9b19c5601b850a6065396ca9809789afe2ea569e65c5a42e48bb26c5bb8d84"} Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.104075 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.210942 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kmgb\" (UniqueName: \"kubernetes.io/projected/0b6cd940-c79b-4bd7-80ff-69bb850709ae-kube-api-access-7kmgb\") pod \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.211074 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-ssh-key\") pod \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.211193 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-inventory\") pod \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\" (UID: \"0b6cd940-c79b-4bd7-80ff-69bb850709ae\") " Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.219027 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6cd940-c79b-4bd7-80ff-69bb850709ae-kube-api-access-7kmgb" (OuterVolumeSpecName: "kube-api-access-7kmgb") pod "0b6cd940-c79b-4bd7-80ff-69bb850709ae" (UID: "0b6cd940-c79b-4bd7-80ff-69bb850709ae"). InnerVolumeSpecName "kube-api-access-7kmgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.244967 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-inventory" (OuterVolumeSpecName: "inventory") pod "0b6cd940-c79b-4bd7-80ff-69bb850709ae" (UID: "0b6cd940-c79b-4bd7-80ff-69bb850709ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.251167 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b6cd940-c79b-4bd7-80ff-69bb850709ae" (UID: "0b6cd940-c79b-4bd7-80ff-69bb850709ae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.314578 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.314633 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b6cd940-c79b-4bd7-80ff-69bb850709ae-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.314647 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kmgb\" (UniqueName: \"kubernetes.io/projected/0b6cd940-c79b-4bd7-80ff-69bb850709ae-kube-api-access-7kmgb\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.687404 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" event={"ID":"0b6cd940-c79b-4bd7-80ff-69bb850709ae","Type":"ContainerDied","Data":"f93da301de2aa28a68b294223f81cb4f2d78c5a9c60b7a28235c9d74305d0abc"} Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.687469 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f93da301de2aa28a68b294223f81cb4f2d78c5a9c60b7a28235c9d74305d0abc" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.687509 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvgld" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.762314 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m"] Nov 28 15:01:22 crc kubenswrapper[4817]: E1128 15:01:22.763916 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6cd940-c79b-4bd7-80ff-69bb850709ae" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.763941 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6cd940-c79b-4bd7-80ff-69bb850709ae" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.764160 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6cd940-c79b-4bd7-80ff-69bb850709ae" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.764799 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.767523 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.767674 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.768097 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.768181 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.777071 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m"] Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.933282 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqtlb\" (UniqueName: \"kubernetes.io/projected/c36cc3fe-b1aa-4858-8c72-1fa223352e09-kube-api-access-dqtlb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.933348 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:22 crc kubenswrapper[4817]: I1128 15:01:22.934079 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.036245 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.036383 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqtlb\" (UniqueName: \"kubernetes.io/projected/c36cc3fe-b1aa-4858-8c72-1fa223352e09-kube-api-access-dqtlb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.036425 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.040006 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.045144 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.058286 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqtlb\" (UniqueName: \"kubernetes.io/projected/c36cc3fe-b1aa-4858-8c72-1fa223352e09-kube-api-access-dqtlb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.143564 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:23 crc kubenswrapper[4817]: I1128 15:01:23.694536 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m"] Nov 28 15:01:24 crc kubenswrapper[4817]: I1128 15:01:24.707741 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" event={"ID":"c36cc3fe-b1aa-4858-8c72-1fa223352e09","Type":"ContainerStarted","Data":"8f8965438beac64b147a0674210eecb465d20d6eb03f43d1fc7eda50c852c4b6"} Nov 28 15:01:24 crc kubenswrapper[4817]: I1128 15:01:24.738015 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:01:25 crc kubenswrapper[4817]: I1128 15:01:25.723629 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"299a0b9a4b4228f050653c1264821b02cbeb7c9099edfe3028175a4a0fc9c023"} Nov 28 15:01:28 crc kubenswrapper[4817]: I1128 15:01:28.752499 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" event={"ID":"c36cc3fe-b1aa-4858-8c72-1fa223352e09","Type":"ContainerStarted","Data":"7be87f052f19e0dbdfbd2649d7ac25863895eb0897438b88e507b293314738e6"} Nov 28 15:01:28 crc kubenswrapper[4817]: I1128 15:01:28.770953 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" podStartSLOduration=2.612699462 podStartE2EDuration="6.770931651s" podCreationTimestamp="2025-11-28 15:01:22 +0000 UTC" firstStartedPulling="2025-11-28 15:01:23.701206177 +0000 UTC m=+1906.289184463" lastFinishedPulling="2025-11-28 15:01:27.859438386 +0000 UTC m=+1910.447416652" observedRunningTime="2025-11-28 15:01:28.770383168 +0000 UTC m=+1911.358361444" watchObservedRunningTime="2025-11-28 15:01:28.770931651 +0000 UTC m=+1911.358909937" Nov 28 15:01:37 crc kubenswrapper[4817]: I1128 15:01:37.842018 4817 generic.go:334] "Generic (PLEG): container finished" podID="c36cc3fe-b1aa-4858-8c72-1fa223352e09" containerID="7be87f052f19e0dbdfbd2649d7ac25863895eb0897438b88e507b293314738e6" exitCode=0 Nov 28 15:01:37 crc kubenswrapper[4817]: I1128 15:01:37.842559 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" event={"ID":"c36cc3fe-b1aa-4858-8c72-1fa223352e09","Type":"ContainerDied","Data":"7be87f052f19e0dbdfbd2649d7ac25863895eb0897438b88e507b293314738e6"} Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.306985 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.381674 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqtlb\" (UniqueName: \"kubernetes.io/projected/c36cc3fe-b1aa-4858-8c72-1fa223352e09-kube-api-access-dqtlb\") pod \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.382094 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-ssh-key\") pod \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.382228 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-inventory\") pod \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\" (UID: \"c36cc3fe-b1aa-4858-8c72-1fa223352e09\") " Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.389428 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c36cc3fe-b1aa-4858-8c72-1fa223352e09-kube-api-access-dqtlb" (OuterVolumeSpecName: "kube-api-access-dqtlb") pod "c36cc3fe-b1aa-4858-8c72-1fa223352e09" (UID: "c36cc3fe-b1aa-4858-8c72-1fa223352e09"). InnerVolumeSpecName "kube-api-access-dqtlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.415100 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-inventory" (OuterVolumeSpecName: "inventory") pod "c36cc3fe-b1aa-4858-8c72-1fa223352e09" (UID: "c36cc3fe-b1aa-4858-8c72-1fa223352e09"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.415131 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c36cc3fe-b1aa-4858-8c72-1fa223352e09" (UID: "c36cc3fe-b1aa-4858-8c72-1fa223352e09"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.484864 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.484930 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqtlb\" (UniqueName: \"kubernetes.io/projected/c36cc3fe-b1aa-4858-8c72-1fa223352e09-kube-api-access-dqtlb\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.484961 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c36cc3fe-b1aa-4858-8c72-1fa223352e09-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.875257 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" event={"ID":"c36cc3fe-b1aa-4858-8c72-1fa223352e09","Type":"ContainerDied","Data":"8f8965438beac64b147a0674210eecb465d20d6eb03f43d1fc7eda50c852c4b6"} Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.875306 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f8965438beac64b147a0674210eecb465d20d6eb03f43d1fc7eda50c852c4b6" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.875592 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.971309 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd"] Nov 28 15:01:39 crc kubenswrapper[4817]: E1128 15:01:39.971809 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36cc3fe-b1aa-4858-8c72-1fa223352e09" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.971828 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36cc3fe-b1aa-4858-8c72-1fa223352e09" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.972062 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c36cc3fe-b1aa-4858-8c72-1fa223352e09" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.972800 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.977190 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.977838 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.978073 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.978546 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.978848 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.978996 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.979158 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.979340 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:01:39 crc kubenswrapper[4817]: I1128 15:01:39.984596 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd"] Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095385 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095764 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095792 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095813 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095872 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095898 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.095944 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096073 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096159 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096228 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096433 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096593 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096676 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxhhc\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-kube-api-access-kxhhc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.096833 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198293 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198374 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198424 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198514 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198613 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198676 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxhhc\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-kube-api-access-kxhhc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198762 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198809 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.198858 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.199017 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.199063 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.199479 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.199577 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.199618 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.205592 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.207058 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.207197 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.207614 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.208440 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.209607 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.209904 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.210935 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.211494 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.213189 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.216680 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.220761 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.221500 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxhhc\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-kube-api-access-kxhhc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.222852 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.321163 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.842619 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd"] Nov 28 15:01:40 crc kubenswrapper[4817]: W1128 15:01:40.844450 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc817bef2_9ff5_4f69_8a73_1f36128879a6.slice/crio-4d7bbc183973858fa0321d7062f06315aa473fc6f4dfc5e6c5d931dd23fbaeeb WatchSource:0}: Error finding container 4d7bbc183973858fa0321d7062f06315aa473fc6f4dfc5e6c5d931dd23fbaeeb: Status 404 returned error can't find the container with id 4d7bbc183973858fa0321d7062f06315aa473fc6f4dfc5e6c5d931dd23fbaeeb Nov 28 15:01:40 crc kubenswrapper[4817]: I1128 15:01:40.890088 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" event={"ID":"c817bef2-9ff5-4f69-8a73-1f36128879a6","Type":"ContainerStarted","Data":"4d7bbc183973858fa0321d7062f06315aa473fc6f4dfc5e6c5d931dd23fbaeeb"} Nov 28 15:01:41 crc kubenswrapper[4817]: I1128 15:01:41.900840 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" event={"ID":"c817bef2-9ff5-4f69-8a73-1f36128879a6","Type":"ContainerStarted","Data":"5774d83861a7236486a4e62eb319d3dbaf9a0ab88c1a337581cdfac6bbaabf03"} Nov 28 15:01:41 crc kubenswrapper[4817]: I1128 15:01:41.932095 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" podStartSLOduration=2.442791648 podStartE2EDuration="2.932072196s" podCreationTimestamp="2025-11-28 15:01:39 +0000 UTC" firstStartedPulling="2025-11-28 15:01:40.847291421 +0000 UTC m=+1923.435269697" lastFinishedPulling="2025-11-28 15:01:41.336571979 +0000 UTC m=+1923.924550245" observedRunningTime="2025-11-28 15:01:41.918818906 +0000 UTC m=+1924.506797172" watchObservedRunningTime="2025-11-28 15:01:41.932072196 +0000 UTC m=+1924.520050462" Nov 28 15:02:21 crc kubenswrapper[4817]: I1128 15:02:21.273384 4817 generic.go:334] "Generic (PLEG): container finished" podID="c817bef2-9ff5-4f69-8a73-1f36128879a6" containerID="5774d83861a7236486a4e62eb319d3dbaf9a0ab88c1a337581cdfac6bbaabf03" exitCode=0 Nov 28 15:02:21 crc kubenswrapper[4817]: I1128 15:02:21.273475 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" event={"ID":"c817bef2-9ff5-4f69-8a73-1f36128879a6","Type":"ContainerDied","Data":"5774d83861a7236486a4e62eb319d3dbaf9a0ab88c1a337581cdfac6bbaabf03"} Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.659812 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.807834 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-neutron-metadata-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.807939 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-bootstrap-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.807970 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-libvirt-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808004 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808155 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ovn-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808200 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-telemetry-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808241 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808283 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-inventory\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808322 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808349 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ssh-key\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808879 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-repo-setup-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808903 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxhhc\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-kube-api-access-kxhhc\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808967 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.808994 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-nova-combined-ca-bundle\") pod \"c817bef2-9ff5-4f69-8a73-1f36128879a6\" (UID: \"c817bef2-9ff5-4f69-8a73-1f36128879a6\") " Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.814339 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.814511 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.814561 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.815383 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.815938 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-kube-api-access-kxhhc" (OuterVolumeSpecName: "kube-api-access-kxhhc") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "kube-api-access-kxhhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.816086 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.816138 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.816689 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.818078 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.818543 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.818940 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.833137 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.845819 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-inventory" (OuterVolumeSpecName: "inventory") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.859144 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c817bef2-9ff5-4f69-8a73-1f36128879a6" (UID: "c817bef2-9ff5-4f69-8a73-1f36128879a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910705 4817 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910749 4817 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910759 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910770 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910780 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910790 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910799 4817 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910809 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxhhc\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-kube-api-access-kxhhc\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910818 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910828 4817 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910837 4817 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910847 4817 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910855 4817 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c817bef2-9ff5-4f69-8a73-1f36128879a6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:22 crc kubenswrapper[4817]: I1128 15:02:22.910867 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c817bef2-9ff5-4f69-8a73-1f36128879a6-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.293944 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" event={"ID":"c817bef2-9ff5-4f69-8a73-1f36128879a6","Type":"ContainerDied","Data":"4d7bbc183973858fa0321d7062f06315aa473fc6f4dfc5e6c5d931dd23fbaeeb"} Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.294191 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d7bbc183973858fa0321d7062f06315aa473fc6f4dfc5e6c5d931dd23fbaeeb" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.294246 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.490408 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6"] Nov 28 15:02:23 crc kubenswrapper[4817]: E1128 15:02:23.490911 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c817bef2-9ff5-4f69-8a73-1f36128879a6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.490930 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c817bef2-9ff5-4f69-8a73-1f36128879a6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.491191 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c817bef2-9ff5-4f69-8a73-1f36128879a6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.492095 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.501037 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.501158 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.502690 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.503131 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.503142 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.508207 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6"] Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.542934 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.543021 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.543078 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.543170 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9mfc\" (UniqueName: \"kubernetes.io/projected/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-kube-api-access-p9mfc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.543215 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.644626 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.644703 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.644805 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9mfc\" (UniqueName: \"kubernetes.io/projected/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-kube-api-access-p9mfc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.644843 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.644881 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.646491 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.649288 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.650478 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.654422 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.666534 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9mfc\" (UniqueName: \"kubernetes.io/projected/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-kube-api-access-p9mfc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-d6ls6\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:23 crc kubenswrapper[4817]: I1128 15:02:23.807647 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:02:24 crc kubenswrapper[4817]: I1128 15:02:24.362347 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6"] Nov 28 15:02:25 crc kubenswrapper[4817]: I1128 15:02:25.310462 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" event={"ID":"c49884f3-a2a8-49d5-8d6c-c2d7485909ed","Type":"ContainerStarted","Data":"eccf05c25c7cb7e1f5c0b3a3f423429b940adab5ee01cc485d5b521152d2f227"} Nov 28 15:02:29 crc kubenswrapper[4817]: I1128 15:02:29.372104 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" event={"ID":"c49884f3-a2a8-49d5-8d6c-c2d7485909ed","Type":"ContainerStarted","Data":"724f4bae6da492c8e1c5b9aee0c213f28a4c1ff9ffe94308fbdc7e5c284a7aa7"} Nov 28 15:02:29 crc kubenswrapper[4817]: I1128 15:02:29.391688 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" podStartSLOduration=2.185239614 podStartE2EDuration="6.391670303s" podCreationTimestamp="2025-11-28 15:02:23 +0000 UTC" firstStartedPulling="2025-11-28 15:02:24.362996732 +0000 UTC m=+1966.950975008" lastFinishedPulling="2025-11-28 15:02:28.569427431 +0000 UTC m=+1971.157405697" observedRunningTime="2025-11-28 15:02:29.386198617 +0000 UTC m=+1971.974176883" watchObservedRunningTime="2025-11-28 15:02:29.391670303 +0000 UTC m=+1971.979648569" Nov 28 15:03:34 crc kubenswrapper[4817]: I1128 15:03:34.044535 4817 generic.go:334] "Generic (PLEG): container finished" podID="c49884f3-a2a8-49d5-8d6c-c2d7485909ed" containerID="724f4bae6da492c8e1c5b9aee0c213f28a4c1ff9ffe94308fbdc7e5c284a7aa7" exitCode=0 Nov 28 15:03:34 crc kubenswrapper[4817]: I1128 15:03:34.044699 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" event={"ID":"c49884f3-a2a8-49d5-8d6c-c2d7485909ed","Type":"ContainerDied","Data":"724f4bae6da492c8e1c5b9aee0c213f28a4c1ff9ffe94308fbdc7e5c284a7aa7"} Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.521852 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.701439 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-inventory\") pod \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.702265 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9mfc\" (UniqueName: \"kubernetes.io/projected/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-kube-api-access-p9mfc\") pod \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.702316 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovn-combined-ca-bundle\") pod \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.702434 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ssh-key\") pod \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.702524 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovncontroller-config-0\") pod \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\" (UID: \"c49884f3-a2a8-49d5-8d6c-c2d7485909ed\") " Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.715629 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c49884f3-a2a8-49d5-8d6c-c2d7485909ed" (UID: "c49884f3-a2a8-49d5-8d6c-c2d7485909ed"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.716889 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-kube-api-access-p9mfc" (OuterVolumeSpecName: "kube-api-access-p9mfc") pod "c49884f3-a2a8-49d5-8d6c-c2d7485909ed" (UID: "c49884f3-a2a8-49d5-8d6c-c2d7485909ed"). InnerVolumeSpecName "kube-api-access-p9mfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.740756 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-inventory" (OuterVolumeSpecName: "inventory") pod "c49884f3-a2a8-49d5-8d6c-c2d7485909ed" (UID: "c49884f3-a2a8-49d5-8d6c-c2d7485909ed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.741280 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c49884f3-a2a8-49d5-8d6c-c2d7485909ed" (UID: "c49884f3-a2a8-49d5-8d6c-c2d7485909ed"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.761774 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c49884f3-a2a8-49d5-8d6c-c2d7485909ed" (UID: "c49884f3-a2a8-49d5-8d6c-c2d7485909ed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.807937 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.807981 4817 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.808002 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9mfc\" (UniqueName: \"kubernetes.io/projected/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-kube-api-access-p9mfc\") on node \"crc\" DevicePath \"\"" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.808019 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:03:35 crc kubenswrapper[4817]: I1128 15:03:35.808060 4817 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c49884f3-a2a8-49d5-8d6c-c2d7485909ed-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.080808 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" event={"ID":"c49884f3-a2a8-49d5-8d6c-c2d7485909ed","Type":"ContainerDied","Data":"eccf05c25c7cb7e1f5c0b3a3f423429b940adab5ee01cc485d5b521152d2f227"} Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.080853 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eccf05c25c7cb7e1f5c0b3a3f423429b940adab5ee01cc485d5b521152d2f227" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.080915 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-d6ls6" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.199536 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c"] Nov 28 15:03:36 crc kubenswrapper[4817]: E1128 15:03:36.200175 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49884f3-a2a8-49d5-8d6c-c2d7485909ed" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.200270 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49884f3-a2a8-49d5-8d6c-c2d7485909ed" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.200488 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c49884f3-a2a8-49d5-8d6c-c2d7485909ed" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.201220 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.208305 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.208647 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.208739 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.209701 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.209888 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.210113 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.214854 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.214942 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjzvt\" (UniqueName: \"kubernetes.io/projected/f7494198-71fe-4460-a0a5-0737e7fe57dd-kube-api-access-wjzvt\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.214995 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.215059 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.215099 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.215170 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.218572 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c"] Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.317184 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.317263 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.317294 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.317341 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.317386 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.317448 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjzvt\" (UniqueName: \"kubernetes.io/projected/f7494198-71fe-4460-a0a5-0737e7fe57dd-kube-api-access-wjzvt\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.320648 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.320704 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.320900 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.321750 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.322232 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.333505 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjzvt\" (UniqueName: \"kubernetes.io/projected/f7494198-71fe-4460-a0a5-0737e7fe57dd-kube-api-access-wjzvt\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:36 crc kubenswrapper[4817]: I1128 15:03:36.522016 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:03:37 crc kubenswrapper[4817]: I1128 15:03:37.046276 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c"] Nov 28 15:03:37 crc kubenswrapper[4817]: I1128 15:03:37.091750 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" event={"ID":"f7494198-71fe-4460-a0a5-0737e7fe57dd","Type":"ContainerStarted","Data":"e8392f0f5511db14e72fc6cf6dc3ff7ebeb4e806280e562718d6d4ebd479412a"} Nov 28 15:03:38 crc kubenswrapper[4817]: I1128 15:03:38.101233 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" event={"ID":"f7494198-71fe-4460-a0a5-0737e7fe57dd","Type":"ContainerStarted","Data":"ca68eb33e9c201f7069d5101a99d011d4430efda87969bdaaa49e0a169931b74"} Nov 28 15:03:54 crc kubenswrapper[4817]: I1128 15:03:54.045441 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:03:54 crc kubenswrapper[4817]: I1128 15:03:54.046144 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:04:24 crc kubenswrapper[4817]: I1128 15:04:24.045242 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:04:24 crc kubenswrapper[4817]: I1128 15:04:24.045862 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:04:26 crc kubenswrapper[4817]: I1128 15:04:26.579084 4817 generic.go:334] "Generic (PLEG): container finished" podID="f7494198-71fe-4460-a0a5-0737e7fe57dd" containerID="ca68eb33e9c201f7069d5101a99d011d4430efda87969bdaaa49e0a169931b74" exitCode=0 Nov 28 15:04:26 crc kubenswrapper[4817]: I1128 15:04:26.579159 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" event={"ID":"f7494198-71fe-4460-a0a5-0737e7fe57dd","Type":"ContainerDied","Data":"ca68eb33e9c201f7069d5101a99d011d4430efda87969bdaaa49e0a169931b74"} Nov 28 15:04:27 crc kubenswrapper[4817]: I1128 15:04:27.956997 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.128053 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"f7494198-71fe-4460-a0a5-0737e7fe57dd\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.128116 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-metadata-combined-ca-bundle\") pod \"f7494198-71fe-4460-a0a5-0737e7fe57dd\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.128220 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjzvt\" (UniqueName: \"kubernetes.io/projected/f7494198-71fe-4460-a0a5-0737e7fe57dd-kube-api-access-wjzvt\") pod \"f7494198-71fe-4460-a0a5-0737e7fe57dd\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.128285 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-nova-metadata-neutron-config-0\") pod \"f7494198-71fe-4460-a0a5-0737e7fe57dd\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.128392 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-inventory\") pod \"f7494198-71fe-4460-a0a5-0737e7fe57dd\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.128419 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-ssh-key\") pod \"f7494198-71fe-4460-a0a5-0737e7fe57dd\" (UID: \"f7494198-71fe-4460-a0a5-0737e7fe57dd\") " Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.133372 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7494198-71fe-4460-a0a5-0737e7fe57dd-kube-api-access-wjzvt" (OuterVolumeSpecName: "kube-api-access-wjzvt") pod "f7494198-71fe-4460-a0a5-0737e7fe57dd" (UID: "f7494198-71fe-4460-a0a5-0737e7fe57dd"). InnerVolumeSpecName "kube-api-access-wjzvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.134046 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f7494198-71fe-4460-a0a5-0737e7fe57dd" (UID: "f7494198-71fe-4460-a0a5-0737e7fe57dd"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.156879 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "f7494198-71fe-4460-a0a5-0737e7fe57dd" (UID: "f7494198-71fe-4460-a0a5-0737e7fe57dd"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.169410 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f7494198-71fe-4460-a0a5-0737e7fe57dd" (UID: "f7494198-71fe-4460-a0a5-0737e7fe57dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.169957 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "f7494198-71fe-4460-a0a5-0737e7fe57dd" (UID: "f7494198-71fe-4460-a0a5-0737e7fe57dd"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.171598 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-inventory" (OuterVolumeSpecName: "inventory") pod "f7494198-71fe-4460-a0a5-0737e7fe57dd" (UID: "f7494198-71fe-4460-a0a5-0737e7fe57dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.231600 4817 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.231825 4817 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.231904 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjzvt\" (UniqueName: \"kubernetes.io/projected/f7494198-71fe-4460-a0a5-0737e7fe57dd-kube-api-access-wjzvt\") on node \"crc\" DevicePath \"\"" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.231975 4817 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.232048 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.232159 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7494198-71fe-4460-a0a5-0737e7fe57dd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.600206 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" event={"ID":"f7494198-71fe-4460-a0a5-0737e7fe57dd","Type":"ContainerDied","Data":"e8392f0f5511db14e72fc6cf6dc3ff7ebeb4e806280e562718d6d4ebd479412a"} Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.600248 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8392f0f5511db14e72fc6cf6dc3ff7ebeb4e806280e562718d6d4ebd479412a" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.600304 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.694937 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm"] Nov 28 15:04:28 crc kubenswrapper[4817]: E1128 15:04:28.695670 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7494198-71fe-4460-a0a5-0737e7fe57dd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.695812 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7494198-71fe-4460-a0a5-0737e7fe57dd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.696129 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7494198-71fe-4460-a0a5-0737e7fe57dd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.696998 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.702760 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.702958 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.703160 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.703190 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.703441 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.705792 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm"] Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.844469 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.844814 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjlxf\" (UniqueName: \"kubernetes.io/projected/58e1259c-59ec-4bee-867b-b90c25667b33-kube-api-access-wjlxf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.844892 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.845033 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.845103 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.946666 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjlxf\" (UniqueName: \"kubernetes.io/projected/58e1259c-59ec-4bee-867b-b90c25667b33-kube-api-access-wjlxf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.946765 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.946807 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.946841 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.946985 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.957360 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.960250 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.960467 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.963660 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:28 crc kubenswrapper[4817]: I1128 15:04:28.973603 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjlxf\" (UniqueName: \"kubernetes.io/projected/58e1259c-59ec-4bee-867b-b90c25667b33-kube-api-access-wjlxf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:29 crc kubenswrapper[4817]: I1128 15:04:29.024310 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:04:29 crc kubenswrapper[4817]: I1128 15:04:29.638607 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm"] Nov 28 15:04:30 crc kubenswrapper[4817]: I1128 15:04:30.621191 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" event={"ID":"58e1259c-59ec-4bee-867b-b90c25667b33","Type":"ContainerStarted","Data":"a3179c1ee7258c0bbd1bf9aac9207481dcf87c0262688c05066c03adc161996e"} Nov 28 15:04:31 crc kubenswrapper[4817]: I1128 15:04:31.630612 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" event={"ID":"58e1259c-59ec-4bee-867b-b90c25667b33","Type":"ContainerStarted","Data":"8f1f7ee8759aa42c1dc446e9ad623f8b49c626d0dd6ef6c8542d3d378f9af995"} Nov 28 15:04:31 crc kubenswrapper[4817]: I1128 15:04:31.659153 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" podStartSLOduration=2.432552675 podStartE2EDuration="3.659135168s" podCreationTimestamp="2025-11-28 15:04:28 +0000 UTC" firstStartedPulling="2025-11-28 15:04:29.628686042 +0000 UTC m=+2092.216664328" lastFinishedPulling="2025-11-28 15:04:30.855268555 +0000 UTC m=+2093.443246821" observedRunningTime="2025-11-28 15:04:31.656192764 +0000 UTC m=+2094.244171030" watchObservedRunningTime="2025-11-28 15:04:31.659135168 +0000 UTC m=+2094.247113434" Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.046082 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.046939 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.047024 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.048324 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"299a0b9a4b4228f050653c1264821b02cbeb7c9099edfe3028175a4a0fc9c023"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.048428 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://299a0b9a4b4228f050653c1264821b02cbeb7c9099edfe3028175a4a0fc9c023" gracePeriod=600 Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.881853 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="299a0b9a4b4228f050653c1264821b02cbeb7c9099edfe3028175a4a0fc9c023" exitCode=0 Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.881970 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"299a0b9a4b4228f050653c1264821b02cbeb7c9099edfe3028175a4a0fc9c023"} Nov 28 15:04:54 crc kubenswrapper[4817]: I1128 15:04:54.882315 4817 scope.go:117] "RemoveContainer" containerID="16c7931b9b97e30250fbd251a17d7d589a96af1e5c4cefd66efd4e2635b1d290" Nov 28 15:04:56 crc kubenswrapper[4817]: I1128 15:04:56.907291 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe"} Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.019049 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-79msl"] Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.023692 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.028674 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-79msl"] Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.085754 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-catalog-content\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.085839 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-utilities\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.085882 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7tnx\" (UniqueName: \"kubernetes.io/projected/69a761ea-1149-46a2-84b8-6c86ad17a659-kube-api-access-d7tnx\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.187147 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7tnx\" (UniqueName: \"kubernetes.io/projected/69a761ea-1149-46a2-84b8-6c86ad17a659-kube-api-access-d7tnx\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.187282 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-catalog-content\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.187325 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-utilities\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.187812 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-utilities\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.188276 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-catalog-content\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.206128 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7tnx\" (UniqueName: \"kubernetes.io/projected/69a761ea-1149-46a2-84b8-6c86ad17a659-kube-api-access-d7tnx\") pod \"certified-operators-79msl\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.354078 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:11 crc kubenswrapper[4817]: I1128 15:07:11.893794 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-79msl"] Nov 28 15:07:12 crc kubenswrapper[4817]: I1128 15:07:12.222805 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79msl" event={"ID":"69a761ea-1149-46a2-84b8-6c86ad17a659","Type":"ContainerStarted","Data":"5a0a7d2546695aa8bd1b441f09c4c46741ab93a2423b2b356a090ffa5b34bfaf"} Nov 28 15:07:13 crc kubenswrapper[4817]: I1128 15:07:13.234735 4817 generic.go:334] "Generic (PLEG): container finished" podID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerID="224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c" exitCode=0 Nov 28 15:07:13 crc kubenswrapper[4817]: I1128 15:07:13.234808 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79msl" event={"ID":"69a761ea-1149-46a2-84b8-6c86ad17a659","Type":"ContainerDied","Data":"224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c"} Nov 28 15:07:13 crc kubenswrapper[4817]: I1128 15:07:13.237184 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:07:16 crc kubenswrapper[4817]: I1128 15:07:16.263117 4817 generic.go:334] "Generic (PLEG): container finished" podID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerID="7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f" exitCode=0 Nov 28 15:07:16 crc kubenswrapper[4817]: I1128 15:07:16.263213 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79msl" event={"ID":"69a761ea-1149-46a2-84b8-6c86ad17a659","Type":"ContainerDied","Data":"7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f"} Nov 28 15:07:18 crc kubenswrapper[4817]: I1128 15:07:18.280912 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79msl" event={"ID":"69a761ea-1149-46a2-84b8-6c86ad17a659","Type":"ContainerStarted","Data":"826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671"} Nov 28 15:07:18 crc kubenswrapper[4817]: I1128 15:07:18.306515 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-79msl" podStartSLOduration=4.258985664 podStartE2EDuration="8.306496124s" podCreationTimestamp="2025-11-28 15:07:10 +0000 UTC" firstStartedPulling="2025-11-28 15:07:13.23699224 +0000 UTC m=+2255.824970506" lastFinishedPulling="2025-11-28 15:07:17.2845027 +0000 UTC m=+2259.872480966" observedRunningTime="2025-11-28 15:07:18.301040517 +0000 UTC m=+2260.889018793" watchObservedRunningTime="2025-11-28 15:07:18.306496124 +0000 UTC m=+2260.894474390" Nov 28 15:07:21 crc kubenswrapper[4817]: I1128 15:07:21.354658 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:21 crc kubenswrapper[4817]: I1128 15:07:21.355062 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:21 crc kubenswrapper[4817]: I1128 15:07:21.402254 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:22 crc kubenswrapper[4817]: I1128 15:07:22.382149 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:22 crc kubenswrapper[4817]: I1128 15:07:22.439198 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-79msl"] Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.045069 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.045363 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.348192 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-79msl" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="registry-server" containerID="cri-o://826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671" gracePeriod=2 Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.821569 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.832677 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-utilities\") pod \"69a761ea-1149-46a2-84b8-6c86ad17a659\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.834106 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-utilities" (OuterVolumeSpecName: "utilities") pod "69a761ea-1149-46a2-84b8-6c86ad17a659" (UID: "69a761ea-1149-46a2-84b8-6c86ad17a659"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.934769 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-catalog-content\") pod \"69a761ea-1149-46a2-84b8-6c86ad17a659\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.934870 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7tnx\" (UniqueName: \"kubernetes.io/projected/69a761ea-1149-46a2-84b8-6c86ad17a659-kube-api-access-d7tnx\") pod \"69a761ea-1149-46a2-84b8-6c86ad17a659\" (UID: \"69a761ea-1149-46a2-84b8-6c86ad17a659\") " Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.935507 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.943948 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69a761ea-1149-46a2-84b8-6c86ad17a659-kube-api-access-d7tnx" (OuterVolumeSpecName: "kube-api-access-d7tnx") pod "69a761ea-1149-46a2-84b8-6c86ad17a659" (UID: "69a761ea-1149-46a2-84b8-6c86ad17a659"). InnerVolumeSpecName "kube-api-access-d7tnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:07:24 crc kubenswrapper[4817]: I1128 15:07:24.995508 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69a761ea-1149-46a2-84b8-6c86ad17a659" (UID: "69a761ea-1149-46a2-84b8-6c86ad17a659"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.036986 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7tnx\" (UniqueName: \"kubernetes.io/projected/69a761ea-1149-46a2-84b8-6c86ad17a659-kube-api-access-d7tnx\") on node \"crc\" DevicePath \"\"" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.037021 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69a761ea-1149-46a2-84b8-6c86ad17a659-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.361149 4817 generic.go:334] "Generic (PLEG): container finished" podID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerID="826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671" exitCode=0 Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.361204 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79msl" event={"ID":"69a761ea-1149-46a2-84b8-6c86ad17a659","Type":"ContainerDied","Data":"826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671"} Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.361238 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-79msl" event={"ID":"69a761ea-1149-46a2-84b8-6c86ad17a659","Type":"ContainerDied","Data":"5a0a7d2546695aa8bd1b441f09c4c46741ab93a2423b2b356a090ffa5b34bfaf"} Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.361261 4817 scope.go:117] "RemoveContainer" containerID="826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.361283 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-79msl" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.414562 4817 scope.go:117] "RemoveContainer" containerID="7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.415848 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-79msl"] Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.425936 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-79msl"] Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.434300 4817 scope.go:117] "RemoveContainer" containerID="224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.491767 4817 scope.go:117] "RemoveContainer" containerID="826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671" Nov 28 15:07:25 crc kubenswrapper[4817]: E1128 15:07:25.492294 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671\": container with ID starting with 826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671 not found: ID does not exist" containerID="826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.492441 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671"} err="failed to get container status \"826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671\": rpc error: code = NotFound desc = could not find container \"826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671\": container with ID starting with 826591856b420ec1563e7cd5a4ccf45d335a7aed83b1c44f9f064191c93d7671 not found: ID does not exist" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.492550 4817 scope.go:117] "RemoveContainer" containerID="7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f" Nov 28 15:07:25 crc kubenswrapper[4817]: E1128 15:07:25.492934 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f\": container with ID starting with 7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f not found: ID does not exist" containerID="7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.492971 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f"} err="failed to get container status \"7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f\": rpc error: code = NotFound desc = could not find container \"7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f\": container with ID starting with 7d503eab1e9d9052d6cb478123d0cc5014086ea7da1091cdbeb16d608827f68f not found: ID does not exist" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.492994 4817 scope.go:117] "RemoveContainer" containerID="224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c" Nov 28 15:07:25 crc kubenswrapper[4817]: E1128 15:07:25.493231 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c\": container with ID starting with 224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c not found: ID does not exist" containerID="224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.493254 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c"} err="failed to get container status \"224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c\": rpc error: code = NotFound desc = could not find container \"224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c\": container with ID starting with 224d3fc4158de7fac956ce0f2c80844520d07b92cd89aa23e65a69bb20bc397c not found: ID does not exist" Nov 28 15:07:25 crc kubenswrapper[4817]: I1128 15:07:25.747822 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" path="/var/lib/kubelet/pods/69a761ea-1149-46a2-84b8-6c86ad17a659/volumes" Nov 28 15:07:54 crc kubenswrapper[4817]: I1128 15:07:54.045814 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:07:54 crc kubenswrapper[4817]: I1128 15:07:54.046493 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.045111 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.045828 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.045872 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.047286 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.047343 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" gracePeriod=600 Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.970443 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" exitCode=0 Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.970511 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe"} Nov 28 15:08:24 crc kubenswrapper[4817]: I1128 15:08:24.970753 4817 scope.go:117] "RemoveContainer" containerID="299a0b9a4b4228f050653c1264821b02cbeb7c9099edfe3028175a4a0fc9c023" Nov 28 15:08:25 crc kubenswrapper[4817]: E1128 15:08:25.460989 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:08:25 crc kubenswrapper[4817]: I1128 15:08:25.984234 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:08:25 crc kubenswrapper[4817]: E1128 15:08:25.984486 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:08:37 crc kubenswrapper[4817]: I1128 15:08:37.742657 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:08:37 crc kubenswrapper[4817]: E1128 15:08:37.743991 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:08:48 crc kubenswrapper[4817]: I1128 15:08:48.222308 4817 generic.go:334] "Generic (PLEG): container finished" podID="58e1259c-59ec-4bee-867b-b90c25667b33" containerID="8f1f7ee8759aa42c1dc446e9ad623f8b49c626d0dd6ef6c8542d3d378f9af995" exitCode=0 Nov 28 15:08:48 crc kubenswrapper[4817]: I1128 15:08:48.222388 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" event={"ID":"58e1259c-59ec-4bee-867b-b90c25667b33","Type":"ContainerDied","Data":"8f1f7ee8759aa42c1dc446e9ad623f8b49c626d0dd6ef6c8542d3d378f9af995"} Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.636998 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.747491 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-ssh-key\") pod \"58e1259c-59ec-4bee-867b-b90c25667b33\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.747610 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-inventory\") pod \"58e1259c-59ec-4bee-867b-b90c25667b33\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.747658 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjlxf\" (UniqueName: \"kubernetes.io/projected/58e1259c-59ec-4bee-867b-b90c25667b33-kube-api-access-wjlxf\") pod \"58e1259c-59ec-4bee-867b-b90c25667b33\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.747711 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-combined-ca-bundle\") pod \"58e1259c-59ec-4bee-867b-b90c25667b33\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.748467 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-secret-0\") pod \"58e1259c-59ec-4bee-867b-b90c25667b33\" (UID: \"58e1259c-59ec-4bee-867b-b90c25667b33\") " Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.753934 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e1259c-59ec-4bee-867b-b90c25667b33-kube-api-access-wjlxf" (OuterVolumeSpecName: "kube-api-access-wjlxf") pod "58e1259c-59ec-4bee-867b-b90c25667b33" (UID: "58e1259c-59ec-4bee-867b-b90c25667b33"). InnerVolumeSpecName "kube-api-access-wjlxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.761908 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "58e1259c-59ec-4bee-867b-b90c25667b33" (UID: "58e1259c-59ec-4bee-867b-b90c25667b33"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.776687 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-inventory" (OuterVolumeSpecName: "inventory") pod "58e1259c-59ec-4bee-867b-b90c25667b33" (UID: "58e1259c-59ec-4bee-867b-b90c25667b33"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.777575 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58e1259c-59ec-4bee-867b-b90c25667b33" (UID: "58e1259c-59ec-4bee-867b-b90c25667b33"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.778981 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "58e1259c-59ec-4bee-867b-b90c25667b33" (UID: "58e1259c-59ec-4bee-867b-b90c25667b33"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.851230 4817 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.851966 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.852116 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.852204 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjlxf\" (UniqueName: \"kubernetes.io/projected/58e1259c-59ec-4bee-867b-b90c25667b33-kube-api-access-wjlxf\") on node \"crc\" DevicePath \"\"" Nov 28 15:08:49 crc kubenswrapper[4817]: I1128 15:08:49.852294 4817 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e1259c-59ec-4bee-867b-b90c25667b33-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.243547 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" event={"ID":"58e1259c-59ec-4bee-867b-b90c25667b33","Type":"ContainerDied","Data":"a3179c1ee7258c0bbd1bf9aac9207481dcf87c0262688c05066c03adc161996e"} Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.243588 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3179c1ee7258c0bbd1bf9aac9207481dcf87c0262688c05066c03adc161996e" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.243601 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.369474 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8"] Nov 28 15:08:50 crc kubenswrapper[4817]: E1128 15:08:50.370464 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="extract-content" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.370513 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="extract-content" Nov 28 15:08:50 crc kubenswrapper[4817]: E1128 15:08:50.370533 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="extract-utilities" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.370541 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="extract-utilities" Nov 28 15:08:50 crc kubenswrapper[4817]: E1128 15:08:50.370560 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="registry-server" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.370567 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="registry-server" Nov 28 15:08:50 crc kubenswrapper[4817]: E1128 15:08:50.370612 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e1259c-59ec-4bee-867b-b90c25667b33" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.370621 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e1259c-59ec-4bee-867b-b90c25667b33" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.371011 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="69a761ea-1149-46a2-84b8-6c86ad17a659" containerName="registry-server" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.371194 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e1259c-59ec-4bee-867b-b90c25667b33" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.372296 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.375562 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.376705 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.377799 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.377869 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.378037 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.378596 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.380499 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.389822 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8"] Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.565768 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svcps\" (UniqueName: \"kubernetes.io/projected/8833fd86-9ef4-441d-ada9-bb24bec33153-kube-api-access-svcps\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.566041 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.566060 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.566078 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.566227 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.566517 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.566570 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.567038 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.567121 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.668441 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669183 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svcps\" (UniqueName: \"kubernetes.io/projected/8833fd86-9ef4-441d-ada9-bb24bec33153-kube-api-access-svcps\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669309 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669393 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669471 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669573 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669683 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669784 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.669906 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.671891 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.678853 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.680829 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.680948 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.681486 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.681586 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.681609 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.684606 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.688619 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svcps\" (UniqueName: \"kubernetes.io/projected/8833fd86-9ef4-441d-ada9-bb24bec33153-kube-api-access-svcps\") pod \"nova-edpm-deployment-openstack-edpm-ipam-g4mz8\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:50 crc kubenswrapper[4817]: I1128 15:08:50.692314 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:08:51 crc kubenswrapper[4817]: I1128 15:08:51.003779 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8"] Nov 28 15:08:51 crc kubenswrapper[4817]: I1128 15:08:51.251503 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" event={"ID":"8833fd86-9ef4-441d-ada9-bb24bec33153","Type":"ContainerStarted","Data":"f6240bf46fa2dabe64c089e0d4212e3f9c1eae0d5fd4e6cf9140b9340a0c2deb"} Nov 28 15:08:52 crc kubenswrapper[4817]: I1128 15:08:52.738795 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:08:52 crc kubenswrapper[4817]: E1128 15:08:52.739684 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:08:53 crc kubenswrapper[4817]: I1128 15:08:53.272370 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" event={"ID":"8833fd86-9ef4-441d-ada9-bb24bec33153","Type":"ContainerStarted","Data":"7360c687700c0d9f2e48208238c94ba85d991535fa4a741f2ef4c53d26c230cd"} Nov 28 15:08:53 crc kubenswrapper[4817]: I1128 15:08:53.298420 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" podStartSLOduration=2.335388868 podStartE2EDuration="3.298399793s" podCreationTimestamp="2025-11-28 15:08:50 +0000 UTC" firstStartedPulling="2025-11-28 15:08:51.007963753 +0000 UTC m=+2353.595942019" lastFinishedPulling="2025-11-28 15:08:51.970974658 +0000 UTC m=+2354.558952944" observedRunningTime="2025-11-28 15:08:53.291259734 +0000 UTC m=+2355.879238010" watchObservedRunningTime="2025-11-28 15:08:53.298399793 +0000 UTC m=+2355.886378059" Nov 28 15:09:04 crc kubenswrapper[4817]: I1128 15:09:04.737858 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:09:04 crc kubenswrapper[4817]: E1128 15:09:04.740034 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:09:16 crc kubenswrapper[4817]: I1128 15:09:16.736822 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:09:16 crc kubenswrapper[4817]: E1128 15:09:16.737615 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:09:17 crc kubenswrapper[4817]: I1128 15:09:17.842321 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zk2sc"] Nov 28 15:09:17 crc kubenswrapper[4817]: I1128 15:09:17.848408 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:17 crc kubenswrapper[4817]: I1128 15:09:17.874608 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zk2sc"] Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.037558 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pr4xv"] Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.039902 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4l9q\" (UniqueName: \"kubernetes.io/projected/f4e64437-c9a2-4d6d-b493-9a59220210f3-kube-api-access-l4l9q\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.040132 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-utilities\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.040406 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-catalog-content\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.041256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.057094 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pr4xv"] Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.143433 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-catalog-content\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.143505 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-catalog-content\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.143575 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4l9q\" (UniqueName: \"kubernetes.io/projected/f4e64437-c9a2-4d6d-b493-9a59220210f3-kube-api-access-l4l9q\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.143626 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7csb\" (UniqueName: \"kubernetes.io/projected/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-kube-api-access-n7csb\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.143646 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-utilities\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.143666 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-utilities\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.144041 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-catalog-content\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.144086 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-utilities\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.175342 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4l9q\" (UniqueName: \"kubernetes.io/projected/f4e64437-c9a2-4d6d-b493-9a59220210f3-kube-api-access-l4l9q\") pod \"redhat-operators-zk2sc\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.175786 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.245135 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7csb\" (UniqueName: \"kubernetes.io/projected/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-kube-api-access-n7csb\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.245535 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-utilities\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.245740 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-catalog-content\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.246142 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-utilities\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.246798 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-catalog-content\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.266653 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7csb\" (UniqueName: \"kubernetes.io/projected/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-kube-api-access-n7csb\") pod \"redhat-marketplace-pr4xv\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.365543 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.731234 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zk2sc"] Nov 28 15:09:18 crc kubenswrapper[4817]: I1128 15:09:18.875083 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pr4xv"] Nov 28 15:09:19 crc kubenswrapper[4817]: I1128 15:09:19.608808 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerID="ee030ed4d67e8f479ce58aac6e98547cc5b2062cf884598d3805ae3b12411e60" exitCode=0 Nov 28 15:09:19 crc kubenswrapper[4817]: I1128 15:09:19.608985 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerDied","Data":"ee030ed4d67e8f479ce58aac6e98547cc5b2062cf884598d3805ae3b12411e60"} Nov 28 15:09:19 crc kubenswrapper[4817]: I1128 15:09:19.609234 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerStarted","Data":"fb1648df5d3546053d743ef1289803a62bd9748ef3b2ead374b865bb3568d5a5"} Nov 28 15:09:19 crc kubenswrapper[4817]: I1128 15:09:19.612889 4817 generic.go:334] "Generic (PLEG): container finished" podID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerID="4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f" exitCode=0 Nov 28 15:09:19 crc kubenswrapper[4817]: I1128 15:09:19.612992 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pr4xv" event={"ID":"3a2d0ee5-004e-4246-8c84-554fe1dffcdc","Type":"ContainerDied","Data":"4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f"} Nov 28 15:09:19 crc kubenswrapper[4817]: I1128 15:09:19.613028 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pr4xv" event={"ID":"3a2d0ee5-004e-4246-8c84-554fe1dffcdc","Type":"ContainerStarted","Data":"657f5c03d36ea7d6d460997c7626744b8fbf4a634b94f322508197e6cd243e6d"} Nov 28 15:09:21 crc kubenswrapper[4817]: I1128 15:09:21.640739 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerStarted","Data":"96f4a79b09265746d25f891358ca54cfe708e0ee654d4094c1ef9965176c9b68"} Nov 28 15:09:21 crc kubenswrapper[4817]: I1128 15:09:21.645504 4817 generic.go:334] "Generic (PLEG): container finished" podID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerID="146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d" exitCode=0 Nov 28 15:09:21 crc kubenswrapper[4817]: I1128 15:09:21.645562 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pr4xv" event={"ID":"3a2d0ee5-004e-4246-8c84-554fe1dffcdc","Type":"ContainerDied","Data":"146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d"} Nov 28 15:09:22 crc kubenswrapper[4817]: I1128 15:09:22.661805 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerID="96f4a79b09265746d25f891358ca54cfe708e0ee654d4094c1ef9965176c9b68" exitCode=0 Nov 28 15:09:22 crc kubenswrapper[4817]: I1128 15:09:22.661843 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerDied","Data":"96f4a79b09265746d25f891358ca54cfe708e0ee654d4094c1ef9965176c9b68"} Nov 28 15:09:23 crc kubenswrapper[4817]: I1128 15:09:23.672494 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pr4xv" event={"ID":"3a2d0ee5-004e-4246-8c84-554fe1dffcdc","Type":"ContainerStarted","Data":"113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049"} Nov 28 15:09:23 crc kubenswrapper[4817]: I1128 15:09:23.700423 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pr4xv" podStartSLOduration=2.793430461 podStartE2EDuration="5.700402887s" podCreationTimestamp="2025-11-28 15:09:18 +0000 UTC" firstStartedPulling="2025-11-28 15:09:19.614417813 +0000 UTC m=+2382.202396079" lastFinishedPulling="2025-11-28 15:09:22.521390239 +0000 UTC m=+2385.109368505" observedRunningTime="2025-11-28 15:09:23.696782366 +0000 UTC m=+2386.284760632" watchObservedRunningTime="2025-11-28 15:09:23.700402887 +0000 UTC m=+2386.288381153" Nov 28 15:09:26 crc kubenswrapper[4817]: I1128 15:09:26.699022 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerStarted","Data":"0f0b9a0180849e2ccdbdb986d6cc8054c66d4aa12e77e60b0d4e2e6e40526f50"} Nov 28 15:09:26 crc kubenswrapper[4817]: I1128 15:09:26.725076 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zk2sc" podStartSLOduration=3.213400519 podStartE2EDuration="9.725057241s" podCreationTimestamp="2025-11-28 15:09:17 +0000 UTC" firstStartedPulling="2025-11-28 15:09:19.61071125 +0000 UTC m=+2382.198689516" lastFinishedPulling="2025-11-28 15:09:26.122367972 +0000 UTC m=+2388.710346238" observedRunningTime="2025-11-28 15:09:26.717959683 +0000 UTC m=+2389.305937959" watchObservedRunningTime="2025-11-28 15:09:26.725057241 +0000 UTC m=+2389.313035497" Nov 28 15:09:28 crc kubenswrapper[4817]: I1128 15:09:28.177212 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:28 crc kubenswrapper[4817]: I1128 15:09:28.177297 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:28 crc kubenswrapper[4817]: I1128 15:09:28.366130 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:28 crc kubenswrapper[4817]: I1128 15:09:28.366190 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:28 crc kubenswrapper[4817]: I1128 15:09:28.437845 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:28 crc kubenswrapper[4817]: I1128 15:09:28.777600 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:29 crc kubenswrapper[4817]: I1128 15:09:29.237315 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pr4xv"] Nov 28 15:09:29 crc kubenswrapper[4817]: I1128 15:09:29.259543 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zk2sc" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="registry-server" probeResult="failure" output=< Nov 28 15:09:29 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 15:09:29 crc kubenswrapper[4817]: > Nov 28 15:09:30 crc kubenswrapper[4817]: I1128 15:09:30.733235 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pr4xv" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="registry-server" containerID="cri-o://113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049" gracePeriod=2 Nov 28 15:09:30 crc kubenswrapper[4817]: I1128 15:09:30.737457 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:09:30 crc kubenswrapper[4817]: E1128 15:09:30.737732 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.180292 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.208318 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-catalog-content\") pod \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.208404 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-utilities\") pod \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.208500 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7csb\" (UniqueName: \"kubernetes.io/projected/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-kube-api-access-n7csb\") pod \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\" (UID: \"3a2d0ee5-004e-4246-8c84-554fe1dffcdc\") " Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.209485 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-utilities" (OuterVolumeSpecName: "utilities") pod "3a2d0ee5-004e-4246-8c84-554fe1dffcdc" (UID: "3a2d0ee5-004e-4246-8c84-554fe1dffcdc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.215050 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-kube-api-access-n7csb" (OuterVolumeSpecName: "kube-api-access-n7csb") pod "3a2d0ee5-004e-4246-8c84-554fe1dffcdc" (UID: "3a2d0ee5-004e-4246-8c84-554fe1dffcdc"). InnerVolumeSpecName "kube-api-access-n7csb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.241906 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a2d0ee5-004e-4246-8c84-554fe1dffcdc" (UID: "3a2d0ee5-004e-4246-8c84-554fe1dffcdc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.310829 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7csb\" (UniqueName: \"kubernetes.io/projected/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-kube-api-access-n7csb\") on node \"crc\" DevicePath \"\"" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.310860 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.310870 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2d0ee5-004e-4246-8c84-554fe1dffcdc-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.745183 4817 generic.go:334] "Generic (PLEG): container finished" podID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerID="113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049" exitCode=0 Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.745321 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pr4xv" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.749978 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pr4xv" event={"ID":"3a2d0ee5-004e-4246-8c84-554fe1dffcdc","Type":"ContainerDied","Data":"113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049"} Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.750024 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pr4xv" event={"ID":"3a2d0ee5-004e-4246-8c84-554fe1dffcdc","Type":"ContainerDied","Data":"657f5c03d36ea7d6d460997c7626744b8fbf4a634b94f322508197e6cd243e6d"} Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.750048 4817 scope.go:117] "RemoveContainer" containerID="113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.768355 4817 scope.go:117] "RemoveContainer" containerID="146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.790323 4817 scope.go:117] "RemoveContainer" containerID="4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.792226 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pr4xv"] Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.799569 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pr4xv"] Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.848662 4817 scope.go:117] "RemoveContainer" containerID="113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049" Nov 28 15:09:31 crc kubenswrapper[4817]: E1128 15:09:31.849167 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049\": container with ID starting with 113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049 not found: ID does not exist" containerID="113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.849215 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049"} err="failed to get container status \"113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049\": rpc error: code = NotFound desc = could not find container \"113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049\": container with ID starting with 113379fff18802304c2f74103c0142f75226c27177ca6004fb68bad696a73049 not found: ID does not exist" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.849251 4817 scope.go:117] "RemoveContainer" containerID="146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d" Nov 28 15:09:31 crc kubenswrapper[4817]: E1128 15:09:31.849616 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d\": container with ID starting with 146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d not found: ID does not exist" containerID="146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.849660 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d"} err="failed to get container status \"146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d\": rpc error: code = NotFound desc = could not find container \"146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d\": container with ID starting with 146e57e673f407249025d015d9370da9fb1c1e342e8cb9a46d6086eb7946697d not found: ID does not exist" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.849681 4817 scope.go:117] "RemoveContainer" containerID="4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f" Nov 28 15:09:31 crc kubenswrapper[4817]: E1128 15:09:31.850075 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f\": container with ID starting with 4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f not found: ID does not exist" containerID="4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f" Nov 28 15:09:31 crc kubenswrapper[4817]: I1128 15:09:31.850114 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f"} err="failed to get container status \"4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f\": rpc error: code = NotFound desc = could not find container \"4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f\": container with ID starting with 4bac76fcfe765a4b8cbed304ed85f3d8dbca125cd4f930d8c3d0eced5798ec7f not found: ID does not exist" Nov 28 15:09:33 crc kubenswrapper[4817]: I1128 15:09:33.751181 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" path="/var/lib/kubelet/pods/3a2d0ee5-004e-4246-8c84-554fe1dffcdc/volumes" Nov 28 15:09:38 crc kubenswrapper[4817]: I1128 15:09:38.237193 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:38 crc kubenswrapper[4817]: I1128 15:09:38.303033 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:38 crc kubenswrapper[4817]: I1128 15:09:38.493583 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zk2sc"] Nov 28 15:09:39 crc kubenswrapper[4817]: I1128 15:09:39.814666 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zk2sc" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="registry-server" containerID="cri-o://0f0b9a0180849e2ccdbdb986d6cc8054c66d4aa12e77e60b0d4e2e6e40526f50" gracePeriod=2 Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.824753 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerDied","Data":"0f0b9a0180849e2ccdbdb986d6cc8054c66d4aa12e77e60b0d4e2e6e40526f50"} Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.824762 4817 generic.go:334] "Generic (PLEG): container finished" podID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerID="0f0b9a0180849e2ccdbdb986d6cc8054c66d4aa12e77e60b0d4e2e6e40526f50" exitCode=0 Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.826294 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zk2sc" event={"ID":"f4e64437-c9a2-4d6d-b493-9a59220210f3","Type":"ContainerDied","Data":"fb1648df5d3546053d743ef1289803a62bd9748ef3b2ead374b865bb3568d5a5"} Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.826362 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb1648df5d3546053d743ef1289803a62bd9748ef3b2ead374b865bb3568d5a5" Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.864489 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.991362 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-utilities\") pod \"f4e64437-c9a2-4d6d-b493-9a59220210f3\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.991610 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4l9q\" (UniqueName: \"kubernetes.io/projected/f4e64437-c9a2-4d6d-b493-9a59220210f3-kube-api-access-l4l9q\") pod \"f4e64437-c9a2-4d6d-b493-9a59220210f3\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.991799 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-catalog-content\") pod \"f4e64437-c9a2-4d6d-b493-9a59220210f3\" (UID: \"f4e64437-c9a2-4d6d-b493-9a59220210f3\") " Nov 28 15:09:40 crc kubenswrapper[4817]: I1128 15:09:40.992547 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-utilities" (OuterVolumeSpecName: "utilities") pod "f4e64437-c9a2-4d6d-b493-9a59220210f3" (UID: "f4e64437-c9a2-4d6d-b493-9a59220210f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.003092 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e64437-c9a2-4d6d-b493-9a59220210f3-kube-api-access-l4l9q" (OuterVolumeSpecName: "kube-api-access-l4l9q") pod "f4e64437-c9a2-4d6d-b493-9a59220210f3" (UID: "f4e64437-c9a2-4d6d-b493-9a59220210f3"). InnerVolumeSpecName "kube-api-access-l4l9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.094921 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.094980 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4l9q\" (UniqueName: \"kubernetes.io/projected/f4e64437-c9a2-4d6d-b493-9a59220210f3-kube-api-access-l4l9q\") on node \"crc\" DevicePath \"\"" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.107311 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4e64437-c9a2-4d6d-b493-9a59220210f3" (UID: "f4e64437-c9a2-4d6d-b493-9a59220210f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.196522 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4e64437-c9a2-4d6d-b493-9a59220210f3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.836150 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zk2sc" Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.864512 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zk2sc"] Nov 28 15:09:41 crc kubenswrapper[4817]: I1128 15:09:41.874362 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zk2sc"] Nov 28 15:09:42 crc kubenswrapper[4817]: I1128 15:09:42.737015 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:09:42 crc kubenswrapper[4817]: E1128 15:09:42.737354 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:09:43 crc kubenswrapper[4817]: I1128 15:09:43.748656 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" path="/var/lib/kubelet/pods/f4e64437-c9a2-4d6d-b493-9a59220210f3/volumes" Nov 28 15:09:55 crc kubenswrapper[4817]: I1128 15:09:55.737244 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:09:55 crc kubenswrapper[4817]: E1128 15:09:55.737787 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:10:10 crc kubenswrapper[4817]: I1128 15:10:10.737301 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:10:10 crc kubenswrapper[4817]: E1128 15:10:10.738502 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:10:23 crc kubenswrapper[4817]: I1128 15:10:23.737801 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:10:23 crc kubenswrapper[4817]: E1128 15:10:23.739166 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:10:38 crc kubenswrapper[4817]: I1128 15:10:38.737963 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:10:38 crc kubenswrapper[4817]: E1128 15:10:38.738971 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:10:51 crc kubenswrapper[4817]: I1128 15:10:51.737661 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:10:51 crc kubenswrapper[4817]: E1128 15:10:51.738577 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.936734 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mf8f4"] Nov 28 15:10:59 crc kubenswrapper[4817]: E1128 15:10:59.937698 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="registry-server" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.937712 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="registry-server" Nov 28 15:10:59 crc kubenswrapper[4817]: E1128 15:10:59.937752 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="extract-utilities" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.937761 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="extract-utilities" Nov 28 15:10:59 crc kubenswrapper[4817]: E1128 15:10:59.937776 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="extract-utilities" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.937786 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="extract-utilities" Nov 28 15:10:59 crc kubenswrapper[4817]: E1128 15:10:59.937812 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="extract-content" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.937822 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="extract-content" Nov 28 15:10:59 crc kubenswrapper[4817]: E1128 15:10:59.937838 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="extract-content" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.937847 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="extract-content" Nov 28 15:10:59 crc kubenswrapper[4817]: E1128 15:10:59.937860 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="registry-server" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.937870 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="registry-server" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.938086 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e64437-c9a2-4d6d-b493-9a59220210f3" containerName="registry-server" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.938103 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2d0ee5-004e-4246-8c84-554fe1dffcdc" containerName="registry-server" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.939663 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:10:59 crc kubenswrapper[4817]: I1128 15:10:59.996101 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mf8f4"] Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.079378 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpz7q\" (UniqueName: \"kubernetes.io/projected/538ed71f-af5e-436a-ad90-e30618f6351c-kube-api-access-rpz7q\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.079425 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-catalog-content\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.079772 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-utilities\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.182105 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpz7q\" (UniqueName: \"kubernetes.io/projected/538ed71f-af5e-436a-ad90-e30618f6351c-kube-api-access-rpz7q\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.182157 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-catalog-content\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.182303 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-utilities\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.183059 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-utilities\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.183104 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-catalog-content\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.203322 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpz7q\" (UniqueName: \"kubernetes.io/projected/538ed71f-af5e-436a-ad90-e30618f6351c-kube-api-access-rpz7q\") pod \"community-operators-mf8f4\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.276428 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:00 crc kubenswrapper[4817]: I1128 15:11:00.814592 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mf8f4"] Nov 28 15:11:01 crc kubenswrapper[4817]: I1128 15:11:01.615807 4817 generic.go:334] "Generic (PLEG): container finished" podID="538ed71f-af5e-436a-ad90-e30618f6351c" containerID="bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f" exitCode=0 Nov 28 15:11:01 crc kubenswrapper[4817]: I1128 15:11:01.615886 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerDied","Data":"bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f"} Nov 28 15:11:01 crc kubenswrapper[4817]: I1128 15:11:01.616213 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerStarted","Data":"d6a74ff4125817f5afac678c44ea3578b578747cead3e9d39ea1ce9739705985"} Nov 28 15:11:02 crc kubenswrapper[4817]: I1128 15:11:02.627793 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerStarted","Data":"121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825"} Nov 28 15:11:03 crc kubenswrapper[4817]: I1128 15:11:03.644132 4817 generic.go:334] "Generic (PLEG): container finished" podID="538ed71f-af5e-436a-ad90-e30618f6351c" containerID="121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825" exitCode=0 Nov 28 15:11:03 crc kubenswrapper[4817]: I1128 15:11:03.644208 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerDied","Data":"121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825"} Nov 28 15:11:03 crc kubenswrapper[4817]: I1128 15:11:03.737772 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:11:03 crc kubenswrapper[4817]: E1128 15:11:03.738231 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:11:04 crc kubenswrapper[4817]: I1128 15:11:04.661166 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerStarted","Data":"02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c"} Nov 28 15:11:04 crc kubenswrapper[4817]: I1128 15:11:04.684787 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mf8f4" podStartSLOduration=3.130247422 podStartE2EDuration="5.684764845s" podCreationTimestamp="2025-11-28 15:10:59 +0000 UTC" firstStartedPulling="2025-11-28 15:11:01.61837665 +0000 UTC m=+2484.206354916" lastFinishedPulling="2025-11-28 15:11:04.172894073 +0000 UTC m=+2486.760872339" observedRunningTime="2025-11-28 15:11:04.677571521 +0000 UTC m=+2487.265549807" watchObservedRunningTime="2025-11-28 15:11:04.684764845 +0000 UTC m=+2487.272743121" Nov 28 15:11:10 crc kubenswrapper[4817]: I1128 15:11:10.277686 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:10 crc kubenswrapper[4817]: I1128 15:11:10.278363 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:10 crc kubenswrapper[4817]: I1128 15:11:10.334171 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:10 crc kubenswrapper[4817]: I1128 15:11:10.786673 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:10 crc kubenswrapper[4817]: I1128 15:11:10.841176 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mf8f4"] Nov 28 15:11:12 crc kubenswrapper[4817]: I1128 15:11:12.736578 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mf8f4" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="registry-server" containerID="cri-o://02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c" gracePeriod=2 Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.260284 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.355733 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-utilities\") pod \"538ed71f-af5e-436a-ad90-e30618f6351c\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.355799 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpz7q\" (UniqueName: \"kubernetes.io/projected/538ed71f-af5e-436a-ad90-e30618f6351c-kube-api-access-rpz7q\") pod \"538ed71f-af5e-436a-ad90-e30618f6351c\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.355889 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-catalog-content\") pod \"538ed71f-af5e-436a-ad90-e30618f6351c\" (UID: \"538ed71f-af5e-436a-ad90-e30618f6351c\") " Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.357282 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-utilities" (OuterVolumeSpecName: "utilities") pod "538ed71f-af5e-436a-ad90-e30618f6351c" (UID: "538ed71f-af5e-436a-ad90-e30618f6351c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.362979 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538ed71f-af5e-436a-ad90-e30618f6351c-kube-api-access-rpz7q" (OuterVolumeSpecName: "kube-api-access-rpz7q") pod "538ed71f-af5e-436a-ad90-e30618f6351c" (UID: "538ed71f-af5e-436a-ad90-e30618f6351c"). InnerVolumeSpecName "kube-api-access-rpz7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.417398 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "538ed71f-af5e-436a-ad90-e30618f6351c" (UID: "538ed71f-af5e-436a-ad90-e30618f6351c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.458334 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.458361 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538ed71f-af5e-436a-ad90-e30618f6351c-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.458371 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpz7q\" (UniqueName: \"kubernetes.io/projected/538ed71f-af5e-436a-ad90-e30618f6351c-kube-api-access-rpz7q\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.752093 4817 generic.go:334] "Generic (PLEG): container finished" podID="538ed71f-af5e-436a-ad90-e30618f6351c" containerID="02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c" exitCode=0 Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.752619 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerDied","Data":"02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c"} Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.752917 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mf8f4" event={"ID":"538ed71f-af5e-436a-ad90-e30618f6351c","Type":"ContainerDied","Data":"d6a74ff4125817f5afac678c44ea3578b578747cead3e9d39ea1ce9739705985"} Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.752965 4817 scope.go:117] "RemoveContainer" containerID="02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.753229 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mf8f4" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.792284 4817 scope.go:117] "RemoveContainer" containerID="121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.815403 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mf8f4"] Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.826913 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mf8f4"] Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.827922 4817 scope.go:117] "RemoveContainer" containerID="bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.883084 4817 scope.go:117] "RemoveContainer" containerID="02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c" Nov 28 15:11:13 crc kubenswrapper[4817]: E1128 15:11:13.883630 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c\": container with ID starting with 02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c not found: ID does not exist" containerID="02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.883666 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c"} err="failed to get container status \"02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c\": rpc error: code = NotFound desc = could not find container \"02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c\": container with ID starting with 02bb0f6288f4e630d2bca25b8b23e2d2b677230ad46ab17c1112834d5f758b9c not found: ID does not exist" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.883694 4817 scope.go:117] "RemoveContainer" containerID="121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825" Nov 28 15:11:13 crc kubenswrapper[4817]: E1128 15:11:13.884222 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825\": container with ID starting with 121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825 not found: ID does not exist" containerID="121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.884295 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825"} err="failed to get container status \"121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825\": rpc error: code = NotFound desc = could not find container \"121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825\": container with ID starting with 121fb071e63d4f1ad3261576cc0a11d0e7006540599dbc1d1fd2a517ba733825 not found: ID does not exist" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.884346 4817 scope.go:117] "RemoveContainer" containerID="bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f" Nov 28 15:11:13 crc kubenswrapper[4817]: E1128 15:11:13.884819 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f\": container with ID starting with bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f not found: ID does not exist" containerID="bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f" Nov 28 15:11:13 crc kubenswrapper[4817]: I1128 15:11:13.884872 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f"} err="failed to get container status \"bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f\": rpc error: code = NotFound desc = could not find container \"bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f\": container with ID starting with bfb28f0b9e06e689b03d2cfb3d57290650ae52f766789b06f4a9558b0da3534f not found: ID does not exist" Nov 28 15:11:15 crc kubenswrapper[4817]: I1128 15:11:15.737197 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:11:15 crc kubenswrapper[4817]: E1128 15:11:15.738024 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:11:15 crc kubenswrapper[4817]: I1128 15:11:15.748708 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" path="/var/lib/kubelet/pods/538ed71f-af5e-436a-ad90-e30618f6351c/volumes" Nov 28 15:11:29 crc kubenswrapper[4817]: I1128 15:11:29.738875 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:11:29 crc kubenswrapper[4817]: E1128 15:11:29.739903 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:11:39 crc kubenswrapper[4817]: I1128 15:11:39.070738 4817 generic.go:334] "Generic (PLEG): container finished" podID="8833fd86-9ef4-441d-ada9-bb24bec33153" containerID="7360c687700c0d9f2e48208238c94ba85d991535fa4a741f2ef4c53d26c230cd" exitCode=0 Nov 28 15:11:39 crc kubenswrapper[4817]: I1128 15:11:39.070797 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" event={"ID":"8833fd86-9ef4-441d-ada9-bb24bec33153","Type":"ContainerDied","Data":"7360c687700c0d9f2e48208238c94ba85d991535fa4a741f2ef4c53d26c230cd"} Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.533754 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721688 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-0\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721769 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svcps\" (UniqueName: \"kubernetes.io/projected/8833fd86-9ef4-441d-ada9-bb24bec33153-kube-api-access-svcps\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721792 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-extra-config-0\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721879 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-1\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721926 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-inventory\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721971 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-0\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.721999 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-combined-ca-bundle\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.722090 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-ssh-key\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.722113 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-1\") pod \"8833fd86-9ef4-441d-ada9-bb24bec33153\" (UID: \"8833fd86-9ef4-441d-ada9-bb24bec33153\") " Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.729157 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8833fd86-9ef4-441d-ada9-bb24bec33153-kube-api-access-svcps" (OuterVolumeSpecName: "kube-api-access-svcps") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "kube-api-access-svcps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.739945 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.751174 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.752026 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.754472 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.756944 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-inventory" (OuterVolumeSpecName: "inventory") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.758068 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.759638 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.759812 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8833fd86-9ef4-441d-ada9-bb24bec33153" (UID: "8833fd86-9ef4-441d-ada9-bb24bec33153"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824418 4817 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824452 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svcps\" (UniqueName: \"kubernetes.io/projected/8833fd86-9ef4-441d-ada9-bb24bec33153-kube-api-access-svcps\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824466 4817 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824479 4817 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824490 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824502 4817 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824512 4817 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824522 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:40 crc kubenswrapper[4817]: I1128 15:11:40.824532 4817 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8833fd86-9ef4-441d-ada9-bb24bec33153-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.096176 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" event={"ID":"8833fd86-9ef4-441d-ada9-bb24bec33153","Type":"ContainerDied","Data":"f6240bf46fa2dabe64c089e0d4212e3f9c1eae0d5fd4e6cf9140b9340a0c2deb"} Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.096212 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-g4mz8" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.096226 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6240bf46fa2dabe64c089e0d4212e3f9c1eae0d5fd4e6cf9140b9340a0c2deb" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.216514 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj"] Nov 28 15:11:41 crc kubenswrapper[4817]: E1128 15:11:41.217493 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="registry-server" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.217636 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="registry-server" Nov 28 15:11:41 crc kubenswrapper[4817]: E1128 15:11:41.217758 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="extract-utilities" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.217870 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="extract-utilities" Nov 28 15:11:41 crc kubenswrapper[4817]: E1128 15:11:41.218003 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="extract-content" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.218117 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="extract-content" Nov 28 15:11:41 crc kubenswrapper[4817]: E1128 15:11:41.218220 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8833fd86-9ef4-441d-ada9-bb24bec33153" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.218325 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8833fd86-9ef4-441d-ada9-bb24bec33153" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.218711 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="538ed71f-af5e-436a-ad90-e30618f6351c" containerName="registry-server" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.218937 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8833fd86-9ef4-441d-ada9-bb24bec33153" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.219978 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.222203 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-grbtp" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.222796 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.222997 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.223370 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.223651 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.226342 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj"] Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.333984 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbtcf\" (UniqueName: \"kubernetes.io/projected/f9a2d436-9de0-46c8-8a53-073060648992-kube-api-access-jbtcf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.334253 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.334376 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.334472 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.334591 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.334825 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.334975 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.436818 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.437188 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.437449 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbtcf\" (UniqueName: \"kubernetes.io/projected/f9a2d436-9de0-46c8-8a53-073060648992-kube-api-access-jbtcf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.437622 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.437800 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.437966 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.438124 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.443090 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.443404 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.444309 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.445426 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.454856 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbtcf\" (UniqueName: \"kubernetes.io/projected/f9a2d436-9de0-46c8-8a53-073060648992-kube-api-access-jbtcf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.465331 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.469443 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-622cj\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:41 crc kubenswrapper[4817]: I1128 15:11:41.539291 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:11:42 crc kubenswrapper[4817]: I1128 15:11:42.230774 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj"] Nov 28 15:11:43 crc kubenswrapper[4817]: I1128 15:11:43.132225 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" event={"ID":"f9a2d436-9de0-46c8-8a53-073060648992","Type":"ContainerStarted","Data":"844becdae11bf5aec3fbbb935296f17bc0fc14a6594f6dafb5cda0b41dc71aff"} Nov 28 15:11:43 crc kubenswrapper[4817]: I1128 15:11:43.132495 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" event={"ID":"f9a2d436-9de0-46c8-8a53-073060648992","Type":"ContainerStarted","Data":"e2338b30caad5fe4f6b3e34a28159d258de7fbceff69bcc1bc9edf90539c4b60"} Nov 28 15:11:43 crc kubenswrapper[4817]: I1128 15:11:43.173791 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" podStartSLOduration=1.575010837 podStartE2EDuration="2.173764698s" podCreationTimestamp="2025-11-28 15:11:41 +0000 UTC" firstStartedPulling="2025-11-28 15:11:42.232664425 +0000 UTC m=+2524.820642691" lastFinishedPulling="2025-11-28 15:11:42.831418256 +0000 UTC m=+2525.419396552" observedRunningTime="2025-11-28 15:11:43.158304853 +0000 UTC m=+2525.746283159" watchObservedRunningTime="2025-11-28 15:11:43.173764698 +0000 UTC m=+2525.761743004" Nov 28 15:11:43 crc kubenswrapper[4817]: I1128 15:11:43.737655 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:11:43 crc kubenswrapper[4817]: E1128 15:11:43.738423 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:11:57 crc kubenswrapper[4817]: I1128 15:11:57.746425 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:11:57 crc kubenswrapper[4817]: E1128 15:11:57.748526 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:12:11 crc kubenswrapper[4817]: I1128 15:12:11.738206 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:12:11 crc kubenswrapper[4817]: E1128 15:12:11.739400 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:12:24 crc kubenswrapper[4817]: I1128 15:12:24.737923 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:12:24 crc kubenswrapper[4817]: E1128 15:12:24.738818 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:12:39 crc kubenswrapper[4817]: I1128 15:12:39.737046 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:12:39 crc kubenswrapper[4817]: E1128 15:12:39.737932 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:12:50 crc kubenswrapper[4817]: I1128 15:12:50.737166 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:12:50 crc kubenswrapper[4817]: E1128 15:12:50.737906 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:13:01 crc kubenswrapper[4817]: I1128 15:13:01.737529 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:13:01 crc kubenswrapper[4817]: E1128 15:13:01.739421 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:13:15 crc kubenswrapper[4817]: I1128 15:13:15.738381 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:13:15 crc kubenswrapper[4817]: E1128 15:13:15.739374 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:13:26 crc kubenswrapper[4817]: I1128 15:13:26.737226 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:13:27 crc kubenswrapper[4817]: I1128 15:13:27.166638 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"ab2a3991476d53fe8266976138f5301f0853c71e1d814b5869f82da9d72b33bf"} Nov 28 15:14:05 crc kubenswrapper[4817]: I1128 15:14:05.595879 4817 generic.go:334] "Generic (PLEG): container finished" podID="f9a2d436-9de0-46c8-8a53-073060648992" containerID="844becdae11bf5aec3fbbb935296f17bc0fc14a6594f6dafb5cda0b41dc71aff" exitCode=0 Nov 28 15:14:05 crc kubenswrapper[4817]: I1128 15:14:05.595960 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" event={"ID":"f9a2d436-9de0-46c8-8a53-073060648992","Type":"ContainerDied","Data":"844becdae11bf5aec3fbbb935296f17bc0fc14a6594f6dafb5cda0b41dc71aff"} Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.116000 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.152070 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbtcf\" (UniqueName: \"kubernetes.io/projected/f9a2d436-9de0-46c8-8a53-073060648992-kube-api-access-jbtcf\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.152395 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-telemetry-combined-ca-bundle\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.152514 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-2\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.152662 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-0\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.152847 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-inventory\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.152999 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ssh-key\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.153177 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-1\") pod \"f9a2d436-9de0-46c8-8a53-073060648992\" (UID: \"f9a2d436-9de0-46c8-8a53-073060648992\") " Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.163527 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.166006 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9a2d436-9de0-46c8-8a53-073060648992-kube-api-access-jbtcf" (OuterVolumeSpecName: "kube-api-access-jbtcf") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "kube-api-access-jbtcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.183678 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-inventory" (OuterVolumeSpecName: "inventory") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.184931 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.193601 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.199964 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.212273 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9a2d436-9de0-46c8-8a53-073060648992" (UID: "f9a2d436-9de0-46c8-8a53-073060648992"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256424 4817 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256476 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbtcf\" (UniqueName: \"kubernetes.io/projected/f9a2d436-9de0-46c8-8a53-073060648992-kube-api-access-jbtcf\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256497 4817 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256515 4817 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256533 4817 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256551 4817 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.256570 4817 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9a2d436-9de0-46c8-8a53-073060648992-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.646961 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" event={"ID":"f9a2d436-9de0-46c8-8a53-073060648992","Type":"ContainerDied","Data":"e2338b30caad5fe4f6b3e34a28159d258de7fbceff69bcc1bc9edf90539c4b60"} Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.647032 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2338b30caad5fe4f6b3e34a28159d258de7fbceff69bcc1bc9edf90539c4b60" Nov 28 15:14:07 crc kubenswrapper[4817]: I1128 15:14:07.647043 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-622cj" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.170337 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx"] Nov 28 15:15:00 crc kubenswrapper[4817]: E1128 15:15:00.171914 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9a2d436-9de0-46c8-8a53-073060648992" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.171947 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9a2d436-9de0-46c8-8a53-073060648992" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.172426 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9a2d436-9de0-46c8-8a53-073060648992" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.173339 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.175286 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.175679 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.198939 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx"] Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.377253 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-config-volume\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.377417 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r47pn\" (UniqueName: \"kubernetes.io/projected/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-kube-api-access-r47pn\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.377492 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-secret-volume\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.479122 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r47pn\" (UniqueName: \"kubernetes.io/projected/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-kube-api-access-r47pn\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.479196 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-secret-volume\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.479369 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-config-volume\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.480245 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-config-volume\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.485734 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-secret-volume\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.495512 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r47pn\" (UniqueName: \"kubernetes.io/projected/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-kube-api-access-r47pn\") pod \"collect-profiles-29405715-hjpcx\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.503141 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:00 crc kubenswrapper[4817]: I1128 15:15:00.949378 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx"] Nov 28 15:15:01 crc kubenswrapper[4817]: I1128 15:15:01.237364 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" event={"ID":"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec","Type":"ContainerStarted","Data":"beca2ba820ca929a3c738d397fdb01bbfddade70cc91afeea8c60a65564b2bae"} Nov 28 15:15:01 crc kubenswrapper[4817]: I1128 15:15:01.237691 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" event={"ID":"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec","Type":"ContainerStarted","Data":"cb9d2be257bba61a24b6ce7ec4d63662f530d6d8070d752aa95b9b3744acb3ae"} Nov 28 15:15:01 crc kubenswrapper[4817]: I1128 15:15:01.260660 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" podStartSLOduration=1.2606363489999999 podStartE2EDuration="1.260636349s" podCreationTimestamp="2025-11-28 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:15:01.253129286 +0000 UTC m=+2723.841107552" watchObservedRunningTime="2025-11-28 15:15:01.260636349 +0000 UTC m=+2723.848614615" Nov 28 15:15:02 crc kubenswrapper[4817]: I1128 15:15:02.250306 4817 generic.go:334] "Generic (PLEG): container finished" podID="1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" containerID="beca2ba820ca929a3c738d397fdb01bbfddade70cc91afeea8c60a65564b2bae" exitCode=0 Nov 28 15:15:02 crc kubenswrapper[4817]: I1128 15:15:02.250420 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" event={"ID":"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec","Type":"ContainerDied","Data":"beca2ba820ca929a3c738d397fdb01bbfddade70cc91afeea8c60a65564b2bae"} Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.601984 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.744060 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-config-volume\") pod \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.744176 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r47pn\" (UniqueName: \"kubernetes.io/projected/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-kube-api-access-r47pn\") pod \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.744379 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-secret-volume\") pod \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\" (UID: \"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec\") " Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.744583 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" (UID: "1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.744820 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.749894 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" (UID: "1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.749920 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-kube-api-access-r47pn" (OuterVolumeSpecName: "kube-api-access-r47pn") pod "1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" (UID: "1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec"). InnerVolumeSpecName "kube-api-access-r47pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.846437 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r47pn\" (UniqueName: \"kubernetes.io/projected/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-kube-api-access-r47pn\") on node \"crc\" DevicePath \"\"" Nov 28 15:15:03 crc kubenswrapper[4817]: I1128 15:15:03.846477 4817 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:15:04 crc kubenswrapper[4817]: I1128 15:15:04.273404 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" event={"ID":"1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec","Type":"ContainerDied","Data":"cb9d2be257bba61a24b6ce7ec4d63662f530d6d8070d752aa95b9b3744acb3ae"} Nov 28 15:15:04 crc kubenswrapper[4817]: I1128 15:15:04.273458 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9d2be257bba61a24b6ce7ec4d63662f530d6d8070d752aa95b9b3744acb3ae" Nov 28 15:15:04 crc kubenswrapper[4817]: I1128 15:15:04.273535 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405715-hjpcx" Nov 28 15:15:04 crc kubenswrapper[4817]: I1128 15:15:04.370052 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx"] Nov 28 15:15:04 crc kubenswrapper[4817]: I1128 15:15:04.382611 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-krkpx"] Nov 28 15:15:05 crc kubenswrapper[4817]: I1128 15:15:05.748701 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d52d05b7-f5b8-4fb7-bd09-e063e48c615d" path="/var/lib/kubelet/pods/d52d05b7-f5b8-4fb7-bd09-e063e48c615d/volumes" Nov 28 15:15:13 crc kubenswrapper[4817]: I1128 15:15:13.830693 4817 scope.go:117] "RemoveContainer" containerID="0b9c1851322fe0c25a724c89367a391dd434a84f03b8746661f38a9d155720b2" Nov 28 15:15:54 crc kubenswrapper[4817]: I1128 15:15:54.045415 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:15:54 crc kubenswrapper[4817]: I1128 15:15:54.046087 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:16:13 crc kubenswrapper[4817]: I1128 15:16:13.906771 4817 scope.go:117] "RemoveContainer" containerID="0f0b9a0180849e2ccdbdb986d6cc8054c66d4aa12e77e60b0d4e2e6e40526f50" Nov 28 15:16:13 crc kubenswrapper[4817]: I1128 15:16:13.945461 4817 scope.go:117] "RemoveContainer" containerID="96f4a79b09265746d25f891358ca54cfe708e0ee654d4094c1ef9965176c9b68" Nov 28 15:16:13 crc kubenswrapper[4817]: I1128 15:16:13.990441 4817 scope.go:117] "RemoveContainer" containerID="ee030ed4d67e8f479ce58aac6e98547cc5b2062cf884598d3805ae3b12411e60" Nov 28 15:16:24 crc kubenswrapper[4817]: I1128 15:16:24.045902 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:16:24 crc kubenswrapper[4817]: I1128 15:16:24.046679 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:16:53 crc kubenswrapper[4817]: I1128 15:16:53.973814 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.044993 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.045084 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.045155 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.046315 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab2a3991476d53fe8266976138f5301f0853c71e1d814b5869f82da9d72b33bf"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.046451 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://ab2a3991476d53fe8266976138f5301f0853c71e1d814b5869f82da9d72b33bf" gracePeriod=600 Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.441675 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="ab2a3991476d53fe8266976138f5301f0853c71e1d814b5869f82da9d72b33bf" exitCode=0 Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.441803 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"ab2a3991476d53fe8266976138f5301f0853c71e1d814b5869f82da9d72b33bf"} Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.442201 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c"} Nov 28 15:16:54 crc kubenswrapper[4817]: I1128 15:16:54.442235 4817 scope.go:117] "RemoveContainer" containerID="627b560ab3c1bddd5e260b45dfd1946afcc957e389895e348ba368a17d0dcebe" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.721564 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.722394 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" containerName="openstackclient" containerID="cri-o://99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa" gracePeriod=2 Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.732632 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.772327 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: E1128 15:16:55.773431 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" containerName="openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.773456 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" containerName="openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: E1128 15:16:55.773487 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" containerName="collect-profiles" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.773496 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" containerName="collect-profiles" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.773734 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d3a1e68-5ee8-4bfc-96b7-745d9a68f6ec" containerName="collect-profiles" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.773767 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" containerName="openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.775387 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.787972 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.798664 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.804405 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.838070 4817 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1b63b3d-750d-49b2-8994-568b7e95ca46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:16:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T15:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\\\",\\\"name\\\":\\\"combined-ca-bundle\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwfzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T15:16:55Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" is forbidden: node \"crc\" can only update pod status for pods with spec.nodeName set to itself" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.849309 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.850579 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.854711 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c1b63b3d-750d-49b2-8994-568b7e95ca46" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" Nov 28 15:16:55 crc kubenswrapper[4817]: E1128 15:16:55.863272 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-vwfzr openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-vwfzr openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="c1b63b3d-750d-49b2-8994-568b7e95ca46" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.866466 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.937845 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config-secret\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.937960 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmc2g\" (UniqueName: \"kubernetes.io/projected/efe3c20f-1ca9-46ef-92cd-de439ba78250-kube-api-access-dmc2g\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.937985 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:55 crc kubenswrapper[4817]: I1128 15:16:55.938011 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-combined-ca-bundle\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.040182 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config-secret\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.040324 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmc2g\" (UniqueName: \"kubernetes.io/projected/efe3c20f-1ca9-46ef-92cd-de439ba78250-kube-api-access-dmc2g\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.040354 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.040387 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-combined-ca-bundle\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.041346 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.047660 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config-secret\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.049583 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-combined-ca-bundle\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.058076 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmc2g\" (UniqueName: \"kubernetes.io/projected/efe3c20f-1ca9-46ef-92cd-de439ba78250-kube-api-access-dmc2g\") pod \"openstackclient\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.179515 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.470623 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.476072 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c1b63b3d-750d-49b2-8994-568b7e95ca46" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.491915 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.496273 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c1b63b3d-750d-49b2-8994-568b7e95ca46" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.717373 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:16:56 crc kubenswrapper[4817]: W1128 15:16:56.721627 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefe3c20f_1ca9_46ef_92cd_de439ba78250.slice/crio-774037f344be10ce5df523e6e539e98ada3d3757aae68e109fcc2876bafdcf7b WatchSource:0}: Error finding container 774037f344be10ce5df523e6e539e98ada3d3757aae68e109fcc2876bafdcf7b: Status 404 returned error can't find the container with id 774037f344be10ce5df523e6e539e98ada3d3757aae68e109fcc2876bafdcf7b Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.948609 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-bxc9f"] Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.950585 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:56 crc kubenswrapper[4817]: I1128 15:16:56.976016 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-bxc9f"] Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.048011 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-429b-account-create-update-vdxjt"] Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.049348 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.051803 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.058407 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-operator-scripts\") pod \"aodh-db-create-bxc9f\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.058468 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lkdj\" (UniqueName: \"kubernetes.io/projected/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-kube-api-access-4lkdj\") pod \"aodh-db-create-bxc9f\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.060669 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-429b-account-create-update-vdxjt"] Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.160336 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09854314-9203-4849-a0c8-b3ce751451fc-operator-scripts\") pod \"aodh-429b-account-create-update-vdxjt\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.160423 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-operator-scripts\") pod \"aodh-db-create-bxc9f\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.160467 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lkdj\" (UniqueName: \"kubernetes.io/projected/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-kube-api-access-4lkdj\") pod \"aodh-db-create-bxc9f\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.160563 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5qn6\" (UniqueName: \"kubernetes.io/projected/09854314-9203-4849-a0c8-b3ce751451fc-kube-api-access-v5qn6\") pod \"aodh-429b-account-create-update-vdxjt\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.161378 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-operator-scripts\") pod \"aodh-db-create-bxc9f\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.182632 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lkdj\" (UniqueName: \"kubernetes.io/projected/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-kube-api-access-4lkdj\") pod \"aodh-db-create-bxc9f\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.263099 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09854314-9203-4849-a0c8-b3ce751451fc-operator-scripts\") pod \"aodh-429b-account-create-update-vdxjt\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.263245 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5qn6\" (UniqueName: \"kubernetes.io/projected/09854314-9203-4849-a0c8-b3ce751451fc-kube-api-access-v5qn6\") pod \"aodh-429b-account-create-update-vdxjt\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.264401 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09854314-9203-4849-a0c8-b3ce751451fc-operator-scripts\") pod \"aodh-429b-account-create-update-vdxjt\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.279376 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5qn6\" (UniqueName: \"kubernetes.io/projected/09854314-9203-4849-a0c8-b3ce751451fc-kube-api-access-v5qn6\") pod \"aodh-429b-account-create-update-vdxjt\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.314407 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.400218 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.485676 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.485697 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"efe3c20f-1ca9-46ef-92cd-de439ba78250","Type":"ContainerStarted","Data":"1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b"} Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.485783 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"efe3c20f-1ca9-46ef-92cd-de439ba78250","Type":"ContainerStarted","Data":"774037f344be10ce5df523e6e539e98ada3d3757aae68e109fcc2876bafdcf7b"} Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.517105 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c1b63b3d-750d-49b2-8994-568b7e95ca46" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.518510 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.518237325 podStartE2EDuration="2.518237325s" podCreationTimestamp="2025-11-28 15:16:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:16:57.510044684 +0000 UTC m=+2840.098022950" watchObservedRunningTime="2025-11-28 15:16:57.518237325 +0000 UTC m=+2840.106215601" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.749318 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1b63b3d-750d-49b2-8994-568b7e95ca46" path="/var/lib/kubelet/pods/c1b63b3d-750d-49b2-8994-568b7e95ca46/volumes" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.799770 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-bxc9f"] Nov 28 15:16:57 crc kubenswrapper[4817]: W1128 15:16:57.812834 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc95aca32_0e83_4fb7_9715_b2c86caf6cbe.slice/crio-4a9655618ba79d67d859336c0c6bdfb157a62c6162b17d035a76b90603cb77e6 WatchSource:0}: Error finding container 4a9655618ba79d67d859336c0c6bdfb157a62c6162b17d035a76b90603cb77e6: Status 404 returned error can't find the container with id 4a9655618ba79d67d859336c0c6bdfb157a62c6162b17d035a76b90603cb77e6 Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.889681 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-429b-account-create-update-vdxjt"] Nov 28 15:16:57 crc kubenswrapper[4817]: W1128 15:16:57.893044 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09854314_9203_4849_a0c8_b3ce751451fc.slice/crio-c6908393e4edd810d2ce55dde98c56d761c5d7cdced1a7d1a288a2baff2c26dd WatchSource:0}: Error finding container c6908393e4edd810d2ce55dde98c56d761c5d7cdced1a7d1a288a2baff2c26dd: Status 404 returned error can't find the container with id c6908393e4edd810d2ce55dde98c56d761c5d7cdced1a7d1a288a2baff2c26dd Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.972959 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:57 crc kubenswrapper[4817]: I1128 15:16:57.976054 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.078228 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42t58\" (UniqueName: \"kubernetes.io/projected/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-kube-api-access-42t58\") pod \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.078332 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config\") pod \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.078361 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config-secret\") pod \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.078412 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-combined-ca-bundle\") pod \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\" (UID: \"dfd00f5c-51d7-4e3a-b454-bafc844d0da3\") " Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.098279 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-kube-api-access-42t58" (OuterVolumeSpecName: "kube-api-access-42t58") pod "dfd00f5c-51d7-4e3a-b454-bafc844d0da3" (UID: "dfd00f5c-51d7-4e3a-b454-bafc844d0da3"). InnerVolumeSpecName "kube-api-access-42t58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.117196 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "dfd00f5c-51d7-4e3a-b454-bafc844d0da3" (UID: "dfd00f5c-51d7-4e3a-b454-bafc844d0da3"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.138036 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfd00f5c-51d7-4e3a-b454-bafc844d0da3" (UID: "dfd00f5c-51d7-4e3a-b454-bafc844d0da3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.152580 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "dfd00f5c-51d7-4e3a-b454-bafc844d0da3" (UID: "dfd00f5c-51d7-4e3a-b454-bafc844d0da3"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.181078 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42t58\" (UniqueName: \"kubernetes.io/projected/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-kube-api-access-42t58\") on node \"crc\" DevicePath \"\"" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.181119 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.181133 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.181146 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfd00f5c-51d7-4e3a-b454-bafc844d0da3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.496355 4817 generic.go:334] "Generic (PLEG): container finished" podID="09854314-9203-4849-a0c8-b3ce751451fc" containerID="8557924642bc3e55980a94a71a4a81ddf1925065d35134e84809338cfd0684b0" exitCode=0 Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.496429 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-429b-account-create-update-vdxjt" event={"ID":"09854314-9203-4849-a0c8-b3ce751451fc","Type":"ContainerDied","Data":"8557924642bc3e55980a94a71a4a81ddf1925065d35134e84809338cfd0684b0"} Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.496457 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-429b-account-create-update-vdxjt" event={"ID":"09854314-9203-4849-a0c8-b3ce751451fc","Type":"ContainerStarted","Data":"c6908393e4edd810d2ce55dde98c56d761c5d7cdced1a7d1a288a2baff2c26dd"} Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.498037 4817 generic.go:334] "Generic (PLEG): container finished" podID="c95aca32-0e83-4fb7-9715-b2c86caf6cbe" containerID="3753b95faeedff614bb23f6cb6b99edeb866342c399b2937458ad6ca8ca1f663" exitCode=0 Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.498222 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-bxc9f" event={"ID":"c95aca32-0e83-4fb7-9715-b2c86caf6cbe","Type":"ContainerDied","Data":"3753b95faeedff614bb23f6cb6b99edeb866342c399b2937458ad6ca8ca1f663"} Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.498257 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-bxc9f" event={"ID":"c95aca32-0e83-4fb7-9715-b2c86caf6cbe","Type":"ContainerStarted","Data":"4a9655618ba79d67d859336c0c6bdfb157a62c6162b17d035a76b90603cb77e6"} Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.500079 4817 generic.go:334] "Generic (PLEG): container finished" podID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" containerID="99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa" exitCode=137 Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.500140 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.500187 4817 scope.go:117] "RemoveContainer" containerID="99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.528173 4817 scope.go:117] "RemoveContainer" containerID="99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa" Nov 28 15:16:58 crc kubenswrapper[4817]: E1128 15:16:58.530294 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa\": container with ID starting with 99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa not found: ID does not exist" containerID="99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.530335 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa"} err="failed to get container status \"99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa\": rpc error: code = NotFound desc = could not find container \"99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa\": container with ID starting with 99eb7ee380adf2ca9cbb522a6137df68378f01b27b72607468711923711955fa not found: ID does not exist" Nov 28 15:16:58 crc kubenswrapper[4817]: I1128 15:16:58.535743 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" Nov 28 15:16:59 crc kubenswrapper[4817]: I1128 15:16:59.755897 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfd00f5c-51d7-4e3a-b454-bafc844d0da3" path="/var/lib/kubelet/pods/dfd00f5c-51d7-4e3a-b454-bafc844d0da3/volumes" Nov 28 15:16:59 crc kubenswrapper[4817]: I1128 15:16:59.898286 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bxc9f" Nov 28 15:16:59 crc kubenswrapper[4817]: I1128 15:16:59.907637 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.015477 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09854314-9203-4849-a0c8-b3ce751451fc-operator-scripts\") pod \"09854314-9203-4849-a0c8-b3ce751451fc\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.015551 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5qn6\" (UniqueName: \"kubernetes.io/projected/09854314-9203-4849-a0c8-b3ce751451fc-kube-api-access-v5qn6\") pod \"09854314-9203-4849-a0c8-b3ce751451fc\" (UID: \"09854314-9203-4849-a0c8-b3ce751451fc\") " Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.015585 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lkdj\" (UniqueName: \"kubernetes.io/projected/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-kube-api-access-4lkdj\") pod \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.015845 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-operator-scripts\") pod \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\" (UID: \"c95aca32-0e83-4fb7-9715-b2c86caf6cbe\") " Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.016401 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c95aca32-0e83-4fb7-9715-b2c86caf6cbe" (UID: "c95aca32-0e83-4fb7-9715-b2c86caf6cbe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.016674 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09854314-9203-4849-a0c8-b3ce751451fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09854314-9203-4849-a0c8-b3ce751451fc" (UID: "09854314-9203-4849-a0c8-b3ce751451fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.020553 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-kube-api-access-4lkdj" (OuterVolumeSpecName: "kube-api-access-4lkdj") pod "c95aca32-0e83-4fb7-9715-b2c86caf6cbe" (UID: "c95aca32-0e83-4fb7-9715-b2c86caf6cbe"). InnerVolumeSpecName "kube-api-access-4lkdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.020950 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09854314-9203-4849-a0c8-b3ce751451fc-kube-api-access-v5qn6" (OuterVolumeSpecName: "kube-api-access-v5qn6") pod "09854314-9203-4849-a0c8-b3ce751451fc" (UID: "09854314-9203-4849-a0c8-b3ce751451fc"). InnerVolumeSpecName "kube-api-access-v5qn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.118346 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.118377 4817 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09854314-9203-4849-a0c8-b3ce751451fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.118386 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5qn6\" (UniqueName: \"kubernetes.io/projected/09854314-9203-4849-a0c8-b3ce751451fc-kube-api-access-v5qn6\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.118396 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lkdj\" (UniqueName: \"kubernetes.io/projected/c95aca32-0e83-4fb7-9715-b2c86caf6cbe-kube-api-access-4lkdj\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.522968 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-429b-account-create-update-vdxjt" event={"ID":"09854314-9203-4849-a0c8-b3ce751451fc","Type":"ContainerDied","Data":"c6908393e4edd810d2ce55dde98c56d761c5d7cdced1a7d1a288a2baff2c26dd"} Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.523015 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6908393e4edd810d2ce55dde98c56d761c5d7cdced1a7d1a288a2baff2c26dd" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.523039 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-429b-account-create-update-vdxjt" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.525544 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-bxc9f" event={"ID":"c95aca32-0e83-4fb7-9715-b2c86caf6cbe","Type":"ContainerDied","Data":"4a9655618ba79d67d859336c0c6bdfb157a62c6162b17d035a76b90603cb77e6"} Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.525601 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a9655618ba79d67d859336c0c6bdfb157a62c6162b17d035a76b90603cb77e6" Nov 28 15:17:00 crc kubenswrapper[4817]: I1128 15:17:00.525659 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-bxc9f" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.366912 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-kx4rn"] Nov 28 15:17:02 crc kubenswrapper[4817]: E1128 15:17:02.367664 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09854314-9203-4849-a0c8-b3ce751451fc" containerName="mariadb-account-create-update" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.367677 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="09854314-9203-4849-a0c8-b3ce751451fc" containerName="mariadb-account-create-update" Nov 28 15:17:02 crc kubenswrapper[4817]: E1128 15:17:02.367694 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95aca32-0e83-4fb7-9715-b2c86caf6cbe" containerName="mariadb-database-create" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.367700 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95aca32-0e83-4fb7-9715-b2c86caf6cbe" containerName="mariadb-database-create" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.367892 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c95aca32-0e83-4fb7-9715-b2c86caf6cbe" containerName="mariadb-database-create" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.367915 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="09854314-9203-4849-a0c8-b3ce751451fc" containerName="mariadb-account-create-update" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.368476 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.370880 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.370943 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.371369 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dzj58" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.374343 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.389907 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-kx4rn"] Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.490614 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-config-data\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.490958 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-combined-ca-bundle\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.491224 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-scripts\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.491276 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w6nm\" (UniqueName: \"kubernetes.io/projected/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-kube-api-access-6w6nm\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.593601 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-scripts\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.593704 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w6nm\" (UniqueName: \"kubernetes.io/projected/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-kube-api-access-6w6nm\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.593918 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-config-data\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.594111 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-combined-ca-bundle\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.605782 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-config-data\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.606859 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-scripts\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.607533 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-combined-ca-bundle\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.625013 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w6nm\" (UniqueName: \"kubernetes.io/projected/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-kube-api-access-6w6nm\") pod \"aodh-db-sync-kx4rn\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:02 crc kubenswrapper[4817]: I1128 15:17:02.691885 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:03 crc kubenswrapper[4817]: I1128 15:17:03.158100 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-kx4rn"] Nov 28 15:17:03 crc kubenswrapper[4817]: I1128 15:17:03.165096 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:17:03 crc kubenswrapper[4817]: I1128 15:17:03.555693 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kx4rn" event={"ID":"9ae908e0-75f9-407b-a4dc-6eecb4792d0c","Type":"ContainerStarted","Data":"20d186100913a62cdd673751948e9ae758db47c12f9ed9d8d180c332a0454f56"} Nov 28 15:17:07 crc kubenswrapper[4817]: I1128 15:17:07.598449 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kx4rn" event={"ID":"9ae908e0-75f9-407b-a4dc-6eecb4792d0c","Type":"ContainerStarted","Data":"1fd46706fa32bab4bc161bfda07359787405ab3d923a276f70572282afa6a692"} Nov 28 15:17:07 crc kubenswrapper[4817]: I1128 15:17:07.632023 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-kx4rn" podStartSLOduration=1.793149643 podStartE2EDuration="5.632003632s" podCreationTimestamp="2025-11-28 15:17:02 +0000 UTC" firstStartedPulling="2025-11-28 15:17:03.164847879 +0000 UTC m=+2845.752826145" lastFinishedPulling="2025-11-28 15:17:07.003701868 +0000 UTC m=+2849.591680134" observedRunningTime="2025-11-28 15:17:07.622380254 +0000 UTC m=+2850.210358520" watchObservedRunningTime="2025-11-28 15:17:07.632003632 +0000 UTC m=+2850.219981898" Nov 28 15:17:10 crc kubenswrapper[4817]: I1128 15:17:10.634160 4817 generic.go:334] "Generic (PLEG): container finished" podID="9ae908e0-75f9-407b-a4dc-6eecb4792d0c" containerID="1fd46706fa32bab4bc161bfda07359787405ab3d923a276f70572282afa6a692" exitCode=0 Nov 28 15:17:10 crc kubenswrapper[4817]: I1128 15:17:10.634253 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kx4rn" event={"ID":"9ae908e0-75f9-407b-a4dc-6eecb4792d0c","Type":"ContainerDied","Data":"1fd46706fa32bab4bc161bfda07359787405ab3d923a276f70572282afa6a692"} Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.098548 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.201698 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w6nm\" (UniqueName: \"kubernetes.io/projected/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-kube-api-access-6w6nm\") pod \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.201882 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-config-data\") pod \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.201929 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-scripts\") pod \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.201957 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-combined-ca-bundle\") pod \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\" (UID: \"9ae908e0-75f9-407b-a4dc-6eecb4792d0c\") " Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.209935 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-scripts" (OuterVolumeSpecName: "scripts") pod "9ae908e0-75f9-407b-a4dc-6eecb4792d0c" (UID: "9ae908e0-75f9-407b-a4dc-6eecb4792d0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.212232 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-kube-api-access-6w6nm" (OuterVolumeSpecName: "kube-api-access-6w6nm") pod "9ae908e0-75f9-407b-a4dc-6eecb4792d0c" (UID: "9ae908e0-75f9-407b-a4dc-6eecb4792d0c"). InnerVolumeSpecName "kube-api-access-6w6nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.229891 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ae908e0-75f9-407b-a4dc-6eecb4792d0c" (UID: "9ae908e0-75f9-407b-a4dc-6eecb4792d0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.232337 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-config-data" (OuterVolumeSpecName: "config-data") pod "9ae908e0-75f9-407b-a4dc-6eecb4792d0c" (UID: "9ae908e0-75f9-407b-a4dc-6eecb4792d0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.306528 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w6nm\" (UniqueName: \"kubernetes.io/projected/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-kube-api-access-6w6nm\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.306584 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.306595 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.306606 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae908e0-75f9-407b-a4dc-6eecb4792d0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.664964 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kx4rn" event={"ID":"9ae908e0-75f9-407b-a4dc-6eecb4792d0c","Type":"ContainerDied","Data":"20d186100913a62cdd673751948e9ae758db47c12f9ed9d8d180c332a0454f56"} Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.665012 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20d186100913a62cdd673751948e9ae758db47c12f9ed9d8d180c332a0454f56" Nov 28 15:17:12 crc kubenswrapper[4817]: I1128 15:17:12.665154 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kx4rn" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.938579 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:16 crc kubenswrapper[4817]: E1128 15:17:16.939656 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ae908e0-75f9-407b-a4dc-6eecb4792d0c" containerName="aodh-db-sync" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.939673 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ae908e0-75f9-407b-a4dc-6eecb4792d0c" containerName="aodh-db-sync" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.939970 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ae908e0-75f9-407b-a4dc-6eecb4792d0c" containerName="aodh-db-sync" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.941914 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.944754 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dzj58" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.945369 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.947028 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 28 15:17:16 crc kubenswrapper[4817]: I1128 15:17:16.966910 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.020249 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-scripts\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.020359 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt7rt\" (UniqueName: \"kubernetes.io/projected/6ae53519-4a68-4c8d-8d33-a34762e468b2-kube-api-access-kt7rt\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.020389 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-config-data\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.020419 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.122047 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.122165 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-scripts\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.122239 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt7rt\" (UniqueName: \"kubernetes.io/projected/6ae53519-4a68-4c8d-8d33-a34762e468b2-kube-api-access-kt7rt\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.122268 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-config-data\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.129697 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.132409 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-config-data\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.146830 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-scripts\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.152750 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt7rt\" (UniqueName: \"kubernetes.io/projected/6ae53519-4a68-4c8d-8d33-a34762e468b2-kube-api-access-kt7rt\") pod \"aodh-0\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.264506 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:17:17 crc kubenswrapper[4817]: I1128 15:17:17.762962 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:18 crc kubenswrapper[4817]: I1128 15:17:18.729946 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerStarted","Data":"fec3bef90c50df1494455ee377c32f7ee143a462a71ab35d36d5eed3087ad9da"} Nov 28 15:17:18 crc kubenswrapper[4817]: I1128 15:17:18.730320 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerStarted","Data":"25d397a309a08ae41381c364aeba758147312f04585b9feab9b74711c9eabb65"} Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.604279 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.606021 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-central-agent" containerID="cri-o://87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9" gracePeriod=30 Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.606247 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="proxy-httpd" containerID="cri-o://6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171" gracePeriod=30 Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.606362 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="sg-core" containerID="cri-o://6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99" gracePeriod=30 Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.606419 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-notification-agent" containerID="cri-o://85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52" gracePeriod=30 Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.728686 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.755406 4817 generic.go:334] "Generic (PLEG): container finished" podID="c589da35-8233-46bc-a0da-4c00355281be" containerID="6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99" exitCode=2 Nov 28 15:17:20 crc kubenswrapper[4817]: I1128 15:17:20.755446 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerDied","Data":"6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99"} Nov 28 15:17:21 crc kubenswrapper[4817]: I1128 15:17:21.772280 4817 generic.go:334] "Generic (PLEG): container finished" podID="c589da35-8233-46bc-a0da-4c00355281be" containerID="6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171" exitCode=0 Nov 28 15:17:21 crc kubenswrapper[4817]: I1128 15:17:21.772805 4817 generic.go:334] "Generic (PLEG): container finished" podID="c589da35-8233-46bc-a0da-4c00355281be" containerID="87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9" exitCode=0 Nov 28 15:17:21 crc kubenswrapper[4817]: I1128 15:17:21.772843 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerDied","Data":"6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171"} Nov 28 15:17:21 crc kubenswrapper[4817]: I1128 15:17:21.772868 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerDied","Data":"87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9"} Nov 28 15:17:21 crc kubenswrapper[4817]: I1128 15:17:21.780493 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerStarted","Data":"a6a74b006a9105a56e95fd4b33eb3134bce76f64d535ac5dda17eeebc9000c58"} Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.643117 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.666908 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-log-httpd\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.666977 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-combined-ca-bundle\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667033 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-scripts\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667060 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-config-data\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667156 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpjn6\" (UniqueName: \"kubernetes.io/projected/c589da35-8233-46bc-a0da-4c00355281be-kube-api-access-lpjn6\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667237 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-run-httpd\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667284 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-ceilometer-tls-certs\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667341 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-sg-core-conf-yaml\") pod \"c589da35-8233-46bc-a0da-4c00355281be\" (UID: \"c589da35-8233-46bc-a0da-4c00355281be\") " Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667521 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.667730 4817 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.668254 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.675483 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c589da35-8233-46bc-a0da-4c00355281be-kube-api-access-lpjn6" (OuterVolumeSpecName: "kube-api-access-lpjn6") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "kube-api-access-lpjn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.678090 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-scripts" (OuterVolumeSpecName: "scripts") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.736713 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.757556 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.771623 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpjn6\" (UniqueName: \"kubernetes.io/projected/c589da35-8233-46bc-a0da-4c00355281be-kube-api-access-lpjn6\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.771658 4817 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c589da35-8233-46bc-a0da-4c00355281be-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.771668 4817 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.771677 4817 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.771699 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.785549 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.803372 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-config-data" (OuterVolumeSpecName: "config-data") pod "c589da35-8233-46bc-a0da-4c00355281be" (UID: "c589da35-8233-46bc-a0da-4c00355281be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.805785 4817 generic.go:334] "Generic (PLEG): container finished" podID="c589da35-8233-46bc-a0da-4c00355281be" containerID="85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52" exitCode=0 Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.805815 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerDied","Data":"85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52"} Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.805857 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c589da35-8233-46bc-a0da-4c00355281be","Type":"ContainerDied","Data":"cd35d0ed3d63ed8e55ae788658af0eb7c4305ec67e6c620c37999123ab12cf2b"} Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.805862 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.805878 4817 scope.go:117] "RemoveContainer" containerID="6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.848988 4817 scope.go:117] "RemoveContainer" containerID="6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.876500 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.876534 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c589da35-8233-46bc-a0da-4c00355281be-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.884624 4817 scope.go:117] "RemoveContainer" containerID="85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.886702 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.900063 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908220 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.908692 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-notification-agent" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908704 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-notification-agent" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.908726 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="proxy-httpd" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908734 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="proxy-httpd" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.908754 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-central-agent" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908761 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-central-agent" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.908768 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="sg-core" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908774 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="sg-core" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908936 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-notification-agent" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908950 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="ceilometer-central-agent" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908962 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="proxy-httpd" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.908985 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="c589da35-8233-46bc-a0da-4c00355281be" containerName="sg-core" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.910621 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.917408 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.917600 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.917939 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.918752 4817 scope.go:117] "RemoveContainer" containerID="87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.927210 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.950679 4817 scope.go:117] "RemoveContainer" containerID="6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.951396 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171\": container with ID starting with 6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171 not found: ID does not exist" containerID="6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.951442 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171"} err="failed to get container status \"6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171\": rpc error: code = NotFound desc = could not find container \"6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171\": container with ID starting with 6aa1efcf9c8154a8511977c2efee3e459fbe5eb0b14d766ffd3a838f90a0b171 not found: ID does not exist" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.951470 4817 scope.go:117] "RemoveContainer" containerID="6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.952209 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99\": container with ID starting with 6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99 not found: ID does not exist" containerID="6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.952238 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99"} err="failed to get container status \"6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99\": rpc error: code = NotFound desc = could not find container \"6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99\": container with ID starting with 6eac3e725e4528f4d7ce44dbc9e99f7d80f0e28d6acd7053db74ca94f6a8dc99 not found: ID does not exist" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.952254 4817 scope.go:117] "RemoveContainer" containerID="85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.952537 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52\": container with ID starting with 85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52 not found: ID does not exist" containerID="85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.952563 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52"} err="failed to get container status \"85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52\": rpc error: code = NotFound desc = could not find container \"85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52\": container with ID starting with 85edb604da555fe15b9b56d401b433d99ab29ea05442f109cb35e4cb275f9b52 not found: ID does not exist" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.952578 4817 scope.go:117] "RemoveContainer" containerID="87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9" Nov 28 15:17:24 crc kubenswrapper[4817]: E1128 15:17:24.952857 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9\": container with ID starting with 87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9 not found: ID does not exist" containerID="87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.952885 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9"} err="failed to get container status \"87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9\": rpc error: code = NotFound desc = could not find container \"87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9\": container with ID starting with 87553b26200e257ee63ac9d4686ccbcb4d10f2a713e7aaa74dc1928b0e78d4b9 not found: ID does not exist" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978044 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978152 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f859e2d-de31-407a-86c0-97a3897b1a57-log-httpd\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978188 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8pdh\" (UniqueName: \"kubernetes.io/projected/6f859e2d-de31-407a-86c0-97a3897b1a57-kube-api-access-d8pdh\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978226 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-scripts\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978264 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f859e2d-de31-407a-86c0-97a3897b1a57-run-httpd\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978296 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978366 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-config-data\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:24 crc kubenswrapper[4817]: I1128 15:17:24.978428 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.079928 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-config-data\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080279 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080302 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080351 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f859e2d-de31-407a-86c0-97a3897b1a57-log-httpd\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080375 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pdh\" (UniqueName: \"kubernetes.io/projected/6f859e2d-de31-407a-86c0-97a3897b1a57-kube-api-access-d8pdh\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080398 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-scripts\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080428 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f859e2d-de31-407a-86c0-97a3897b1a57-run-httpd\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.080449 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.081787 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f859e2d-de31-407a-86c0-97a3897b1a57-log-httpd\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.082175 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6f859e2d-de31-407a-86c0-97a3897b1a57-run-httpd\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.086176 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.086230 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-scripts\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.086191 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.086199 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.086781 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f859e2d-de31-407a-86c0-97a3897b1a57-config-data\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.108567 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pdh\" (UniqueName: \"kubernetes.io/projected/6f859e2d-de31-407a-86c0-97a3897b1a57-kube-api-access-d8pdh\") pod \"ceilometer-0\" (UID: \"6f859e2d-de31-407a-86c0-97a3897b1a57\") " pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.238334 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.750984 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c589da35-8233-46bc-a0da-4c00355281be" path="/var/lib/kubelet/pods/c589da35-8233-46bc-a0da-4c00355281be/volumes" Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.788414 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 15:17:25 crc kubenswrapper[4817]: W1128 15:17:25.793455 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f859e2d_de31_407a_86c0_97a3897b1a57.slice/crio-7d6d90d6b8be6b0d18a674a3e44df1dc75842e6d280acedabdc270d5388975d1 WatchSource:0}: Error finding container 7d6d90d6b8be6b0d18a674a3e44df1dc75842e6d280acedabdc270d5388975d1: Status 404 returned error can't find the container with id 7d6d90d6b8be6b0d18a674a3e44df1dc75842e6d280acedabdc270d5388975d1 Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.815504 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f859e2d-de31-407a-86c0-97a3897b1a57","Type":"ContainerStarted","Data":"7d6d90d6b8be6b0d18a674a3e44df1dc75842e6d280acedabdc270d5388975d1"} Nov 28 15:17:25 crc kubenswrapper[4817]: I1128 15:17:25.817474 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerStarted","Data":"e2dac4bba13ffbdf2d6b72564111b748fe3ddae0a3a5c29421d2513b0364a29e"} Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.843924 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f859e2d-de31-407a-86c0-97a3897b1a57","Type":"ContainerStarted","Data":"5cb01cbf048fe7baca713375de4034c529d1f13d64303ff41ee04c98a37e3169"} Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.847239 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerStarted","Data":"1d1c620b67cc9fe68fa6fe415b21c512c8329460607a4e3ecf9303384f441859"} Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.847381 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-api" containerID="cri-o://fec3bef90c50df1494455ee377c32f7ee143a462a71ab35d36d5eed3087ad9da" gracePeriod=30 Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.847447 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-listener" containerID="cri-o://1d1c620b67cc9fe68fa6fe415b21c512c8329460607a4e3ecf9303384f441859" gracePeriod=30 Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.847482 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-evaluator" containerID="cri-o://a6a74b006a9105a56e95fd4b33eb3134bce76f64d535ac5dda17eeebc9000c58" gracePeriod=30 Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.847658 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-notifier" containerID="cri-o://e2dac4bba13ffbdf2d6b72564111b748fe3ddae0a3a5c29421d2513b0364a29e" gracePeriod=30 Nov 28 15:17:27 crc kubenswrapper[4817]: I1128 15:17:27.875331 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.734029214 podStartE2EDuration="11.87531655s" podCreationTimestamp="2025-11-28 15:17:16 +0000 UTC" firstStartedPulling="2025-11-28 15:17:17.771811477 +0000 UTC m=+2860.359789753" lastFinishedPulling="2025-11-28 15:17:26.913098823 +0000 UTC m=+2869.501077089" observedRunningTime="2025-11-28 15:17:27.874209682 +0000 UTC m=+2870.462187948" watchObservedRunningTime="2025-11-28 15:17:27.87531655 +0000 UTC m=+2870.463294816" Nov 28 15:17:28 crc kubenswrapper[4817]: I1128 15:17:28.863301 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f859e2d-de31-407a-86c0-97a3897b1a57","Type":"ContainerStarted","Data":"04128a72d1b194e3a63056c51ff9410ac08988033391b8f3267425a37c98db40"} Nov 28 15:17:28 crc kubenswrapper[4817]: I1128 15:17:28.867068 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerID="a6a74b006a9105a56e95fd4b33eb3134bce76f64d535ac5dda17eeebc9000c58" exitCode=0 Nov 28 15:17:28 crc kubenswrapper[4817]: I1128 15:17:28.867109 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerID="fec3bef90c50df1494455ee377c32f7ee143a462a71ab35d36d5eed3087ad9da" exitCode=0 Nov 28 15:17:28 crc kubenswrapper[4817]: I1128 15:17:28.867134 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerDied","Data":"a6a74b006a9105a56e95fd4b33eb3134bce76f64d535ac5dda17eeebc9000c58"} Nov 28 15:17:28 crc kubenswrapper[4817]: I1128 15:17:28.867162 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerDied","Data":"fec3bef90c50df1494455ee377c32f7ee143a462a71ab35d36d5eed3087ad9da"} Nov 28 15:17:30 crc kubenswrapper[4817]: I1128 15:17:30.890073 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f859e2d-de31-407a-86c0-97a3897b1a57","Type":"ContainerStarted","Data":"11bf01df770809f4b22ba8575fe3bd9cf1a1ecdef04f0b667f05df7dd234770b"} Nov 28 15:17:31 crc kubenswrapper[4817]: I1128 15:17:31.908612 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6f859e2d-de31-407a-86c0-97a3897b1a57","Type":"ContainerStarted","Data":"44885f6aec8f0d6860ddf7198ff54257ba47017cf72d6556058dba728b7ba867"} Nov 28 15:17:31 crc kubenswrapper[4817]: I1128 15:17:31.908946 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 15:17:31 crc kubenswrapper[4817]: I1128 15:17:31.946041 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.440553128 podStartE2EDuration="7.945125391s" podCreationTimestamp="2025-11-28 15:17:24 +0000 UTC" firstStartedPulling="2025-11-28 15:17:25.795912449 +0000 UTC m=+2868.383890725" lastFinishedPulling="2025-11-28 15:17:31.300484702 +0000 UTC m=+2873.888462988" observedRunningTime="2025-11-28 15:17:31.933931615 +0000 UTC m=+2874.521909901" watchObservedRunningTime="2025-11-28 15:17:31.945125391 +0000 UTC m=+2874.533103667" Nov 28 15:17:55 crc kubenswrapper[4817]: I1128 15:17:55.246020 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.183758 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerID="1d1c620b67cc9fe68fa6fe415b21c512c8329460607a4e3ecf9303384f441859" exitCode=137 Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.184370 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerID="e2dac4bba13ffbdf2d6b72564111b748fe3ddae0a3a5c29421d2513b0364a29e" exitCode=137 Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.183845 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerDied","Data":"1d1c620b67cc9fe68fa6fe415b21c512c8329460607a4e3ecf9303384f441859"} Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.184415 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerDied","Data":"e2dac4bba13ffbdf2d6b72564111b748fe3ddae0a3a5c29421d2513b0364a29e"} Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.383559 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.455567 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt7rt\" (UniqueName: \"kubernetes.io/projected/6ae53519-4a68-4c8d-8d33-a34762e468b2-kube-api-access-kt7rt\") pod \"6ae53519-4a68-4c8d-8d33-a34762e468b2\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.456006 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-combined-ca-bundle\") pod \"6ae53519-4a68-4c8d-8d33-a34762e468b2\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.456047 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-config-data\") pod \"6ae53519-4a68-4c8d-8d33-a34762e468b2\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.456070 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-scripts\") pod \"6ae53519-4a68-4c8d-8d33-a34762e468b2\" (UID: \"6ae53519-4a68-4c8d-8d33-a34762e468b2\") " Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.470885 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae53519-4a68-4c8d-8d33-a34762e468b2-kube-api-access-kt7rt" (OuterVolumeSpecName: "kube-api-access-kt7rt") pod "6ae53519-4a68-4c8d-8d33-a34762e468b2" (UID: "6ae53519-4a68-4c8d-8d33-a34762e468b2"). InnerVolumeSpecName "kube-api-access-kt7rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.492954 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-scripts" (OuterVolumeSpecName: "scripts") pod "6ae53519-4a68-4c8d-8d33-a34762e468b2" (UID: "6ae53519-4a68-4c8d-8d33-a34762e468b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.557474 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.557503 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt7rt\" (UniqueName: \"kubernetes.io/projected/6ae53519-4a68-4c8d-8d33-a34762e468b2-kube-api-access-kt7rt\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.609404 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-config-data" (OuterVolumeSpecName: "config-data") pod "6ae53519-4a68-4c8d-8d33-a34762e468b2" (UID: "6ae53519-4a68-4c8d-8d33-a34762e468b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.655390 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ae53519-4a68-4c8d-8d33-a34762e468b2" (UID: "6ae53519-4a68-4c8d-8d33-a34762e468b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.659003 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:58 crc kubenswrapper[4817]: I1128 15:17:58.659028 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae53519-4a68-4c8d-8d33-a34762e468b2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.200575 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6ae53519-4a68-4c8d-8d33-a34762e468b2","Type":"ContainerDied","Data":"25d397a309a08ae41381c364aeba758147312f04585b9feab9b74711c9eabb65"} Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.200656 4817 scope.go:117] "RemoveContainer" containerID="1d1c620b67cc9fe68fa6fe415b21c512c8329460607a4e3ecf9303384f441859" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.200657 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.231876 4817 scope.go:117] "RemoveContainer" containerID="e2dac4bba13ffbdf2d6b72564111b748fe3ddae0a3a5c29421d2513b0364a29e" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.271907 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.289212 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.295473 4817 scope.go:117] "RemoveContainer" containerID="a6a74b006a9105a56e95fd4b33eb3134bce76f64d535ac5dda17eeebc9000c58" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.329547 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.342457 4817 scope.go:117] "RemoveContainer" containerID="fec3bef90c50df1494455ee377c32f7ee143a462a71ab35d36d5eed3087ad9da" Nov 28 15:17:59 crc kubenswrapper[4817]: E1128 15:17:59.346255 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-evaluator" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.347057 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-evaluator" Nov 28 15:17:59 crc kubenswrapper[4817]: E1128 15:17:59.347384 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-api" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.347562 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-api" Nov 28 15:17:59 crc kubenswrapper[4817]: E1128 15:17:59.347909 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-listener" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.348515 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-listener" Nov 28 15:17:59 crc kubenswrapper[4817]: E1128 15:17:59.348988 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-notifier" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.354733 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-notifier" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.355414 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-evaluator" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.355566 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-notifier" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.355651 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-listener" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.355733 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" containerName="aodh-api" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.370014 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.370130 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.373155 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dzj58" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.373336 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.373398 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.373691 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.374700 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-config-data\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.374996 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfv9x\" (UniqueName: \"kubernetes.io/projected/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-kube-api-access-pfv9x\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.375049 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-internal-tls-certs\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.375111 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.375159 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-scripts\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.375180 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-public-tls-certs\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.376257 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.476431 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-config-data\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.476692 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfv9x\" (UniqueName: \"kubernetes.io/projected/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-kube-api-access-pfv9x\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.477130 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-internal-tls-certs\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.477670 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.477925 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-scripts\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.478061 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-public-tls-certs\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.480902 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.481060 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-internal-tls-certs\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.481604 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-scripts\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.481863 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-config-data\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.483558 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-public-tls-certs\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.493984 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfv9x\" (UniqueName: \"kubernetes.io/projected/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-kube-api-access-pfv9x\") pod \"aodh-0\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.689113 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:17:59 crc kubenswrapper[4817]: I1128 15:17:59.750782 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae53519-4a68-4c8d-8d33-a34762e468b2" path="/var/lib/kubelet/pods/6ae53519-4a68-4c8d-8d33-a34762e468b2/volumes" Nov 28 15:18:00 crc kubenswrapper[4817]: I1128 15:18:00.181557 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:18:00 crc kubenswrapper[4817]: W1128 15:18:00.186297 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod332f7b9f_dd4c_4bde_b1d2_3bbb06f8701c.slice/crio-caf819c0569f239e71e207bbaeaefa0bcc5d2cc31224ade3ca24af9cb1d19f02 WatchSource:0}: Error finding container caf819c0569f239e71e207bbaeaefa0bcc5d2cc31224ade3ca24af9cb1d19f02: Status 404 returned error can't find the container with id caf819c0569f239e71e207bbaeaefa0bcc5d2cc31224ade3ca24af9cb1d19f02 Nov 28 15:18:00 crc kubenswrapper[4817]: I1128 15:18:00.214398 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerStarted","Data":"caf819c0569f239e71e207bbaeaefa0bcc5d2cc31224ade3ca24af9cb1d19f02"} Nov 28 15:18:02 crc kubenswrapper[4817]: I1128 15:18:02.237095 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerStarted","Data":"e6342f7d19900580ceafb5ceea16f4586d9af2ea0ebb4304e8d091d131d2d19d"} Nov 28 15:18:03 crc kubenswrapper[4817]: I1128 15:18:03.252641 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerStarted","Data":"8052cf05e1ff2893d30364ce1fd7349dd3c60eccd6e24f7d9ce87cc1249c3bd9"} Nov 28 15:18:03 crc kubenswrapper[4817]: I1128 15:18:03.253392 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerStarted","Data":"2f656a9de3fcca471f43f632b6e9fb00a7e1676c2d800c633566136f16e79d8a"} Nov 28 15:18:04 crc kubenswrapper[4817]: I1128 15:18:04.264933 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerStarted","Data":"f00192d3b73dfcd6b0e6da9485fa02b33d0386b0c9ae8f16ac6e3384a6409b16"} Nov 28 15:18:04 crc kubenswrapper[4817]: I1128 15:18:04.301665 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.782473508 podStartE2EDuration="5.30164163s" podCreationTimestamp="2025-11-28 15:17:59 +0000 UTC" firstStartedPulling="2025-11-28 15:18:00.18880072 +0000 UTC m=+2902.776778986" lastFinishedPulling="2025-11-28 15:18:03.707968822 +0000 UTC m=+2906.295947108" observedRunningTime="2025-11-28 15:18:04.289447868 +0000 UTC m=+2906.877426144" watchObservedRunningTime="2025-11-28 15:18:04.30164163 +0000 UTC m=+2906.889619896" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.629476 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v2lx8"] Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.632893 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.660294 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2lx8"] Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.744182 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-utilities\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.744284 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-catalog-content\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.744388 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rwwq\" (UniqueName: \"kubernetes.io/projected/44617212-204c-478e-8234-6ce08397145f-kube-api-access-4rwwq\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.845900 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-catalog-content\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.846591 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-catalog-content\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.847482 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rwwq\" (UniqueName: \"kubernetes.io/projected/44617212-204c-478e-8234-6ce08397145f-kube-api-access-4rwwq\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.848217 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-utilities\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.848648 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-utilities\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.870218 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rwwq\" (UniqueName: \"kubernetes.io/projected/44617212-204c-478e-8234-6ce08397145f-kube-api-access-4rwwq\") pod \"certified-operators-v2lx8\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:30 crc kubenswrapper[4817]: I1128 15:18:30.978126 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:31 crc kubenswrapper[4817]: I1128 15:18:31.518612 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2lx8"] Nov 28 15:18:31 crc kubenswrapper[4817]: I1128 15:18:31.561368 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerStarted","Data":"bc963e91db5522270bbc02c752a9b31b2459c63e74ef45de28627d4f2ecc3e05"} Nov 28 15:18:32 crc kubenswrapper[4817]: I1128 15:18:32.573315 4817 generic.go:334] "Generic (PLEG): container finished" podID="44617212-204c-478e-8234-6ce08397145f" containerID="d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03" exitCode=0 Nov 28 15:18:32 crc kubenswrapper[4817]: I1128 15:18:32.573426 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerDied","Data":"d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03"} Nov 28 15:18:33 crc kubenswrapper[4817]: I1128 15:18:33.596476 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerStarted","Data":"fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8"} Nov 28 15:18:34 crc kubenswrapper[4817]: I1128 15:18:34.614095 4817 generic.go:334] "Generic (PLEG): container finished" podID="44617212-204c-478e-8234-6ce08397145f" containerID="fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8" exitCode=0 Nov 28 15:18:34 crc kubenswrapper[4817]: I1128 15:18:34.614285 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerDied","Data":"fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8"} Nov 28 15:18:35 crc kubenswrapper[4817]: I1128 15:18:35.627662 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerStarted","Data":"1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231"} Nov 28 15:18:35 crc kubenswrapper[4817]: I1128 15:18:35.649732 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v2lx8" podStartSLOduration=3.033499737 podStartE2EDuration="5.649698344s" podCreationTimestamp="2025-11-28 15:18:30 +0000 UTC" firstStartedPulling="2025-11-28 15:18:32.575922194 +0000 UTC m=+2935.163900460" lastFinishedPulling="2025-11-28 15:18:35.192120781 +0000 UTC m=+2937.780099067" observedRunningTime="2025-11-28 15:18:35.646581014 +0000 UTC m=+2938.234559290" watchObservedRunningTime="2025-11-28 15:18:35.649698344 +0000 UTC m=+2938.237676620" Nov 28 15:18:40 crc kubenswrapper[4817]: I1128 15:18:40.978678 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:40 crc kubenswrapper[4817]: I1128 15:18:40.979388 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:41 crc kubenswrapper[4817]: I1128 15:18:41.057396 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:41 crc kubenswrapper[4817]: I1128 15:18:41.762363 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:43 crc kubenswrapper[4817]: I1128 15:18:43.405899 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2lx8"] Nov 28 15:18:43 crc kubenswrapper[4817]: I1128 15:18:43.721466 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v2lx8" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="registry-server" containerID="cri-o://1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231" gracePeriod=2 Nov 28 15:18:43 crc kubenswrapper[4817]: E1128 15:18:43.908945 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44617212_204c_478e_8234_6ce08397145f.slice/crio-1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44617212_204c_478e_8234_6ce08397145f.slice/crio-conmon-1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.292621 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.451575 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-utilities\") pod \"44617212-204c-478e-8234-6ce08397145f\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.451643 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-catalog-content\") pod \"44617212-204c-478e-8234-6ce08397145f\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.451766 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rwwq\" (UniqueName: \"kubernetes.io/projected/44617212-204c-478e-8234-6ce08397145f-kube-api-access-4rwwq\") pod \"44617212-204c-478e-8234-6ce08397145f\" (UID: \"44617212-204c-478e-8234-6ce08397145f\") " Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.452942 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-utilities" (OuterVolumeSpecName: "utilities") pod "44617212-204c-478e-8234-6ce08397145f" (UID: "44617212-204c-478e-8234-6ce08397145f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.457990 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44617212-204c-478e-8234-6ce08397145f-kube-api-access-4rwwq" (OuterVolumeSpecName: "kube-api-access-4rwwq") pod "44617212-204c-478e-8234-6ce08397145f" (UID: "44617212-204c-478e-8234-6ce08397145f"). InnerVolumeSpecName "kube-api-access-4rwwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.539397 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44617212-204c-478e-8234-6ce08397145f" (UID: "44617212-204c-478e-8234-6ce08397145f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.554251 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.554434 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44617212-204c-478e-8234-6ce08397145f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.554527 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rwwq\" (UniqueName: \"kubernetes.io/projected/44617212-204c-478e-8234-6ce08397145f-kube-api-access-4rwwq\") on node \"crc\" DevicePath \"\"" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.737596 4817 generic.go:334] "Generic (PLEG): container finished" podID="44617212-204c-478e-8234-6ce08397145f" containerID="1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231" exitCode=0 Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.737645 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerDied","Data":"1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231"} Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.737715 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2lx8" event={"ID":"44617212-204c-478e-8234-6ce08397145f","Type":"ContainerDied","Data":"bc963e91db5522270bbc02c752a9b31b2459c63e74ef45de28627d4f2ecc3e05"} Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.737774 4817 scope.go:117] "RemoveContainer" containerID="1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.738600 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2lx8" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.774916 4817 scope.go:117] "RemoveContainer" containerID="fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.800325 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2lx8"] Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.814849 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v2lx8"] Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.828862 4817 scope.go:117] "RemoveContainer" containerID="d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.865348 4817 scope.go:117] "RemoveContainer" containerID="1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231" Nov 28 15:18:44 crc kubenswrapper[4817]: E1128 15:18:44.865861 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231\": container with ID starting with 1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231 not found: ID does not exist" containerID="1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.865891 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231"} err="failed to get container status \"1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231\": rpc error: code = NotFound desc = could not find container \"1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231\": container with ID starting with 1f22534c891d8479696175dcf4516e9acf368d27aaa07c025c3295581b942231 not found: ID does not exist" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.865910 4817 scope.go:117] "RemoveContainer" containerID="fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8" Nov 28 15:18:44 crc kubenswrapper[4817]: E1128 15:18:44.866374 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8\": container with ID starting with fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8 not found: ID does not exist" containerID="fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.866415 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8"} err="failed to get container status \"fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8\": rpc error: code = NotFound desc = could not find container \"fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8\": container with ID starting with fcd4ab87cceebd039c5d52b258ed5c29d8fa3dedbf59490d0cde405daf7ee9e8 not found: ID does not exist" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.866443 4817 scope.go:117] "RemoveContainer" containerID="d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03" Nov 28 15:18:44 crc kubenswrapper[4817]: E1128 15:18:44.866692 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03\": container with ID starting with d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03 not found: ID does not exist" containerID="d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03" Nov 28 15:18:44 crc kubenswrapper[4817]: I1128 15:18:44.866733 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03"} err="failed to get container status \"d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03\": rpc error: code = NotFound desc = could not find container \"d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03\": container with ID starting with d096412f092685bcec5df0bff6122215207f6701bc3caaefa300c5693c06bb03 not found: ID does not exist" Nov 28 15:18:45 crc kubenswrapper[4817]: I1128 15:18:45.758843 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44617212-204c-478e-8234-6ce08397145f" path="/var/lib/kubelet/pods/44617212-204c-478e-8234-6ce08397145f/volumes" Nov 28 15:18:54 crc kubenswrapper[4817]: I1128 15:18:54.044814 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:18:54 crc kubenswrapper[4817]: I1128 15:18:54.045380 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:19:24 crc kubenswrapper[4817]: I1128 15:19:24.045368 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:19:24 crc kubenswrapper[4817]: I1128 15:19:24.046104 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.440244 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mqjkz"] Nov 28 15:19:25 crc kubenswrapper[4817]: E1128 15:19:25.441267 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="registry-server" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.441288 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="registry-server" Nov 28 15:19:25 crc kubenswrapper[4817]: E1128 15:19:25.441302 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="extract-utilities" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.441313 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="extract-utilities" Nov 28 15:19:25 crc kubenswrapper[4817]: E1128 15:19:25.441366 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="extract-content" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.441377 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="extract-content" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.441668 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="44617212-204c-478e-8234-6ce08397145f" containerName="registry-server" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.443575 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.455141 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqjkz"] Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.558339 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-utilities\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.559010 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvh7p\" (UniqueName: \"kubernetes.io/projected/01a07d21-a11f-40ae-9410-2419f7a28820-kube-api-access-cvh7p\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.559113 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-catalog-content\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.661040 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-catalog-content\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.661249 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-utilities\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.661308 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvh7p\" (UniqueName: \"kubernetes.io/projected/01a07d21-a11f-40ae-9410-2419f7a28820-kube-api-access-cvh7p\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.661931 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-catalog-content\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.661985 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-utilities\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.694098 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvh7p\" (UniqueName: \"kubernetes.io/projected/01a07d21-a11f-40ae-9410-2419f7a28820-kube-api-access-cvh7p\") pod \"redhat-operators-mqjkz\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:25 crc kubenswrapper[4817]: I1128 15:19:25.773124 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:26 crc kubenswrapper[4817]: I1128 15:19:26.282024 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqjkz"] Nov 28 15:19:27 crc kubenswrapper[4817]: I1128 15:19:27.214090 4817 generic.go:334] "Generic (PLEG): container finished" podID="01a07d21-a11f-40ae-9410-2419f7a28820" containerID="768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90" exitCode=0 Nov 28 15:19:27 crc kubenswrapper[4817]: I1128 15:19:27.214241 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerDied","Data":"768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90"} Nov 28 15:19:27 crc kubenswrapper[4817]: I1128 15:19:27.214462 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerStarted","Data":"78a6ac225481513ecc3d1a3812412c7a769fd6278f4fd78c9d06a1342a5db6bc"} Nov 28 15:19:29 crc kubenswrapper[4817]: I1128 15:19:29.238445 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerStarted","Data":"90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811"} Nov 28 15:19:30 crc kubenswrapper[4817]: I1128 15:19:30.252045 4817 generic.go:334] "Generic (PLEG): container finished" podID="01a07d21-a11f-40ae-9410-2419f7a28820" containerID="90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811" exitCode=0 Nov 28 15:19:30 crc kubenswrapper[4817]: I1128 15:19:30.252107 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerDied","Data":"90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811"} Nov 28 15:19:44 crc kubenswrapper[4817]: I1128 15:19:44.394338 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerStarted","Data":"d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4"} Nov 28 15:19:44 crc kubenswrapper[4817]: I1128 15:19:44.421268 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mqjkz" podStartSLOduration=2.771165762 podStartE2EDuration="19.42124664s" podCreationTimestamp="2025-11-28 15:19:25 +0000 UTC" firstStartedPulling="2025-11-28 15:19:27.215762593 +0000 UTC m=+2989.803740869" lastFinishedPulling="2025-11-28 15:19:43.865843481 +0000 UTC m=+3006.453821747" observedRunningTime="2025-11-28 15:19:44.412911229 +0000 UTC m=+3007.000889505" watchObservedRunningTime="2025-11-28 15:19:44.42124664 +0000 UTC m=+3007.009224916" Nov 28 15:19:45 crc kubenswrapper[4817]: I1128 15:19:45.773870 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:45 crc kubenswrapper[4817]: I1128 15:19:45.774181 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:46 crc kubenswrapper[4817]: I1128 15:19:46.816820 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mqjkz" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="registry-server" probeResult="failure" output=< Nov 28 15:19:46 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 15:19:46 crc kubenswrapper[4817]: > Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.044801 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.045218 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.045253 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.045861 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.045919 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" gracePeriod=600 Nov 28 15:19:54 crc kubenswrapper[4817]: E1128 15:19:54.168706 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.489216 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" exitCode=0 Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.489261 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c"} Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.489299 4817 scope.go:117] "RemoveContainer" containerID="ab2a3991476d53fe8266976138f5301f0853c71e1d814b5869f82da9d72b33bf" Nov 28 15:19:54 crc kubenswrapper[4817]: I1128 15:19:54.490205 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:19:54 crc kubenswrapper[4817]: E1128 15:19:54.490864 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:19:55 crc kubenswrapper[4817]: I1128 15:19:55.823865 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:55 crc kubenswrapper[4817]: I1128 15:19:55.882455 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:56 crc kubenswrapper[4817]: I1128 15:19:56.663117 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqjkz"] Nov 28 15:19:57 crc kubenswrapper[4817]: I1128 15:19:57.518902 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mqjkz" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="registry-server" containerID="cri-o://d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4" gracePeriod=2 Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.172677 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.268088 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvh7p\" (UniqueName: \"kubernetes.io/projected/01a07d21-a11f-40ae-9410-2419f7a28820-kube-api-access-cvh7p\") pod \"01a07d21-a11f-40ae-9410-2419f7a28820\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.268306 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-catalog-content\") pod \"01a07d21-a11f-40ae-9410-2419f7a28820\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.268438 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-utilities\") pod \"01a07d21-a11f-40ae-9410-2419f7a28820\" (UID: \"01a07d21-a11f-40ae-9410-2419f7a28820\") " Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.269362 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-utilities" (OuterVolumeSpecName: "utilities") pod "01a07d21-a11f-40ae-9410-2419f7a28820" (UID: "01a07d21-a11f-40ae-9410-2419f7a28820"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.273921 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a07d21-a11f-40ae-9410-2419f7a28820-kube-api-access-cvh7p" (OuterVolumeSpecName: "kube-api-access-cvh7p") pod "01a07d21-a11f-40ae-9410-2419f7a28820" (UID: "01a07d21-a11f-40ae-9410-2419f7a28820"). InnerVolumeSpecName "kube-api-access-cvh7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.371099 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.371131 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvh7p\" (UniqueName: \"kubernetes.io/projected/01a07d21-a11f-40ae-9410-2419f7a28820-kube-api-access-cvh7p\") on node \"crc\" DevicePath \"\"" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.389508 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01a07d21-a11f-40ae-9410-2419f7a28820" (UID: "01a07d21-a11f-40ae-9410-2419f7a28820"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.472465 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01a07d21-a11f-40ae-9410-2419f7a28820-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.536184 4817 generic.go:334] "Generic (PLEG): container finished" podID="01a07d21-a11f-40ae-9410-2419f7a28820" containerID="d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4" exitCode=0 Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.536878 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqjkz" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.536935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerDied","Data":"d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4"} Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.537039 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqjkz" event={"ID":"01a07d21-a11f-40ae-9410-2419f7a28820","Type":"ContainerDied","Data":"78a6ac225481513ecc3d1a3812412c7a769fd6278f4fd78c9d06a1342a5db6bc"} Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.537102 4817 scope.go:117] "RemoveContainer" containerID="d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.560821 4817 scope.go:117] "RemoveContainer" containerID="90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.571358 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqjkz"] Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.581895 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mqjkz"] Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.588527 4817 scope.go:117] "RemoveContainer" containerID="768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.642682 4817 scope.go:117] "RemoveContainer" containerID="d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4" Nov 28 15:19:58 crc kubenswrapper[4817]: E1128 15:19:58.643582 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4\": container with ID starting with d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4 not found: ID does not exist" containerID="d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.643629 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4"} err="failed to get container status \"d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4\": rpc error: code = NotFound desc = could not find container \"d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4\": container with ID starting with d0a1ca6f4bcee90d6abcc04f086bd32c43c4da3a795e393216f2f38d3f72afb4 not found: ID does not exist" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.643659 4817 scope.go:117] "RemoveContainer" containerID="90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811" Nov 28 15:19:58 crc kubenswrapper[4817]: E1128 15:19:58.644094 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811\": container with ID starting with 90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811 not found: ID does not exist" containerID="90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.644129 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811"} err="failed to get container status \"90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811\": rpc error: code = NotFound desc = could not find container \"90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811\": container with ID starting with 90af7c0faadd41fbf82e4045a22227c33b7b386c8025780f549ce228f3f19811 not found: ID does not exist" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.644147 4817 scope.go:117] "RemoveContainer" containerID="768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90" Nov 28 15:19:58 crc kubenswrapper[4817]: E1128 15:19:58.644542 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90\": container with ID starting with 768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90 not found: ID does not exist" containerID="768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90" Nov 28 15:19:58 crc kubenswrapper[4817]: I1128 15:19:58.644578 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90"} err="failed to get container status \"768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90\": rpc error: code = NotFound desc = could not find container \"768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90\": container with ID starting with 768eecfbe9a423a8310f25a3d1461f648fd4b8848e0a57596c585f6fe329bc90 not found: ID does not exist" Nov 28 15:19:59 crc kubenswrapper[4817]: I1128 15:19:59.752511 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" path="/var/lib/kubelet/pods/01a07d21-a11f-40ae-9410-2419f7a28820/volumes" Nov 28 15:20:08 crc kubenswrapper[4817]: I1128 15:20:08.737158 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:20:08 crc kubenswrapper[4817]: E1128 15:20:08.738090 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:20:23 crc kubenswrapper[4817]: I1128 15:20:23.737347 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:20:23 crc kubenswrapper[4817]: E1128 15:20:23.738117 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:20:35 crc kubenswrapper[4817]: I1128 15:20:35.737400 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:20:35 crc kubenswrapper[4817]: E1128 15:20:35.738445 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:20:47 crc kubenswrapper[4817]: I1128 15:20:47.751551 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:20:47 crc kubenswrapper[4817]: E1128 15:20:47.752635 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:20:57 crc kubenswrapper[4817]: I1128 15:20:57.133077 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:21:01 crc kubenswrapper[4817]: I1128 15:21:01.737075 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:21:01 crc kubenswrapper[4817]: E1128 15:21:01.738230 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.193268 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz"] Nov 28 15:21:11 crc kubenswrapper[4817]: E1128 15:21:11.194317 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="registry-server" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.194334 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="registry-server" Nov 28 15:21:11 crc kubenswrapper[4817]: E1128 15:21:11.194363 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="extract-content" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.194369 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="extract-content" Nov 28 15:21:11 crc kubenswrapper[4817]: E1128 15:21:11.194386 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="extract-utilities" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.194392 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="extract-utilities" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.194589 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a07d21-a11f-40ae-9410-2419f7a28820" containerName="registry-server" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.195994 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.207564 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.209243 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz"] Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.273859 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.274036 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.274135 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw696\" (UniqueName: \"kubernetes.io/projected/dc724ad5-cdc7-401d-b139-7aa1ff965410-kube-api-access-zw696\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.375708 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.375860 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.375946 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw696\" (UniqueName: \"kubernetes.io/projected/dc724ad5-cdc7-401d-b139-7aa1ff965410-kube-api-access-zw696\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.376625 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.376770 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.404215 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw696\" (UniqueName: \"kubernetes.io/projected/dc724ad5-cdc7-401d-b139-7aa1ff965410-kube-api-access-zw696\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:11 crc kubenswrapper[4817]: I1128 15:21:11.522448 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:12 crc kubenswrapper[4817]: I1128 15:21:12.022200 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz"] Nov 28 15:21:12 crc kubenswrapper[4817]: I1128 15:21:12.342936 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" event={"ID":"dc724ad5-cdc7-401d-b139-7aa1ff965410","Type":"ContainerStarted","Data":"fed5fdbf36c5cb49147a4971713eb02428dd55d7f9568bca2615b26720f3e47c"} Nov 28 15:21:12 crc kubenswrapper[4817]: I1128 15:21:12.343220 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" event={"ID":"dc724ad5-cdc7-401d-b139-7aa1ff965410","Type":"ContainerStarted","Data":"8af657d95bc2ebe320797c6f8326fd2218b693d16ac707e29840154da1412fd6"} Nov 28 15:21:13 crc kubenswrapper[4817]: I1128 15:21:13.354335 4817 generic.go:334] "Generic (PLEG): container finished" podID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerID="fed5fdbf36c5cb49147a4971713eb02428dd55d7f9568bca2615b26720f3e47c" exitCode=0 Nov 28 15:21:13 crc kubenswrapper[4817]: I1128 15:21:13.354385 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" event={"ID":"dc724ad5-cdc7-401d-b139-7aa1ff965410","Type":"ContainerDied","Data":"fed5fdbf36c5cb49147a4971713eb02428dd55d7f9568bca2615b26720f3e47c"} Nov 28 15:21:13 crc kubenswrapper[4817]: I1128 15:21:13.738248 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:21:13 crc kubenswrapper[4817]: E1128 15:21:13.739179 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:21:15 crc kubenswrapper[4817]: I1128 15:21:15.381136 4817 generic.go:334] "Generic (PLEG): container finished" podID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerID="07ef3082b6180bbf8f6b9372e7dd766137d7868a7feb78049748b32eb179538e" exitCode=0 Nov 28 15:21:15 crc kubenswrapper[4817]: I1128 15:21:15.381615 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" event={"ID":"dc724ad5-cdc7-401d-b139-7aa1ff965410","Type":"ContainerDied","Data":"07ef3082b6180bbf8f6b9372e7dd766137d7868a7feb78049748b32eb179538e"} Nov 28 15:21:16 crc kubenswrapper[4817]: I1128 15:21:16.395212 4817 generic.go:334] "Generic (PLEG): container finished" podID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerID="097186df64ffac6ab65e7dc470f353e5042013a30a256c78200b276ff608ff67" exitCode=0 Nov 28 15:21:16 crc kubenswrapper[4817]: I1128 15:21:16.395324 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" event={"ID":"dc724ad5-cdc7-401d-b139-7aa1ff965410","Type":"ContainerDied","Data":"097186df64ffac6ab65e7dc470f353e5042013a30a256c78200b276ff608ff67"} Nov 28 15:21:17 crc kubenswrapper[4817]: I1128 15:21:17.831008 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:17 crc kubenswrapper[4817]: I1128 15:21:17.958263 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-bundle\") pod \"dc724ad5-cdc7-401d-b139-7aa1ff965410\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " Nov 28 15:21:17 crc kubenswrapper[4817]: I1128 15:21:17.958498 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-util\") pod \"dc724ad5-cdc7-401d-b139-7aa1ff965410\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " Nov 28 15:21:17 crc kubenswrapper[4817]: I1128 15:21:17.958541 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw696\" (UniqueName: \"kubernetes.io/projected/dc724ad5-cdc7-401d-b139-7aa1ff965410-kube-api-access-zw696\") pod \"dc724ad5-cdc7-401d-b139-7aa1ff965410\" (UID: \"dc724ad5-cdc7-401d-b139-7aa1ff965410\") " Nov 28 15:21:17 crc kubenswrapper[4817]: I1128 15:21:17.962344 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-bundle" (OuterVolumeSpecName: "bundle") pod "dc724ad5-cdc7-401d-b139-7aa1ff965410" (UID: "dc724ad5-cdc7-401d-b139-7aa1ff965410"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.034354 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc724ad5-cdc7-401d-b139-7aa1ff965410-kube-api-access-zw696" (OuterVolumeSpecName: "kube-api-access-zw696") pod "dc724ad5-cdc7-401d-b139-7aa1ff965410" (UID: "dc724ad5-cdc7-401d-b139-7aa1ff965410"). InnerVolumeSpecName "kube-api-access-zw696". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.062570 4817 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.063163 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw696\" (UniqueName: \"kubernetes.io/projected/dc724ad5-cdc7-401d-b139-7aa1ff965410-kube-api-access-zw696\") on node \"crc\" DevicePath \"\"" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.116706 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-util" (OuterVolumeSpecName: "util") pod "dc724ad5-cdc7-401d-b139-7aa1ff965410" (UID: "dc724ad5-cdc7-401d-b139-7aa1ff965410"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.167018 4817 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc724ad5-cdc7-401d-b139-7aa1ff965410-util\") on node \"crc\" DevicePath \"\"" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.426991 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" event={"ID":"dc724ad5-cdc7-401d-b139-7aa1ff965410","Type":"ContainerDied","Data":"8af657d95bc2ebe320797c6f8326fd2218b693d16ac707e29840154da1412fd6"} Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.427057 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8af657d95bc2ebe320797c6f8326fd2218b693d16ac707e29840154da1412fd6" Nov 28 15:21:18 crc kubenswrapper[4817]: I1128 15:21:18.427110 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz" Nov 28 15:21:26 crc kubenswrapper[4817]: I1128 15:21:26.736556 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:21:26 crc kubenswrapper[4817]: E1128 15:21:26.737231 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.961077 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq"] Nov 28 15:21:28 crc kubenswrapper[4817]: E1128 15:21:28.961682 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="pull" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.961693 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="pull" Nov 28 15:21:28 crc kubenswrapper[4817]: E1128 15:21:28.961711 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="extract" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.961729 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="extract" Nov 28 15:21:28 crc kubenswrapper[4817]: E1128 15:21:28.961750 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="util" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.961757 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="util" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.961958 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc724ad5-cdc7-401d-b139-7aa1ff965410" containerName="extract" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.962853 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.965584 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.967354 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-d9wkv" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.970021 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 28 15:21:28 crc kubenswrapper[4817]: I1128 15:21:28.976509 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.028504 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.030120 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.033927 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-jwdrp" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.034259 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.037177 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.038306 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.051587 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.068540 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.080922 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/597ad435-7283-469d-8c9a-b0e9a032a433-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm\" (UID: \"597ad435-7283-469d-8c9a-b0e9a032a433\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.081160 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/597ad435-7283-469d-8c9a-b0e9a032a433-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm\" (UID: \"597ad435-7283-469d-8c9a-b0e9a032a433\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.081244 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw\" (UID: \"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.081327 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw\" (UID: \"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.081491 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lk47\" (UniqueName: \"kubernetes.io/projected/ad4b53c3-5d16-412b-ba96-f4ba1a54624a-kube-api-access-2lk47\") pod \"obo-prometheus-operator-668cf9dfbb-jvrtq\" (UID: \"ad4b53c3-5d16-412b-ba96-f4ba1a54624a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.183809 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/597ad435-7283-469d-8c9a-b0e9a032a433-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm\" (UID: \"597ad435-7283-469d-8c9a-b0e9a032a433\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.184119 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/597ad435-7283-469d-8c9a-b0e9a032a433-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm\" (UID: \"597ad435-7283-469d-8c9a-b0e9a032a433\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.184153 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw\" (UID: \"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.184190 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw\" (UID: \"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.184220 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lk47\" (UniqueName: \"kubernetes.io/projected/ad4b53c3-5d16-412b-ba96-f4ba1a54624a-kube-api-access-2lk47\") pod \"obo-prometheus-operator-668cf9dfbb-jvrtq\" (UID: \"ad4b53c3-5d16-412b-ba96-f4ba1a54624a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.187051 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-7mc44"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.188350 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.192206 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw\" (UID: \"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.192626 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.192811 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-fjbrh" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.192924 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/597ad435-7283-469d-8c9a-b0e9a032a433-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm\" (UID: \"597ad435-7283-469d-8c9a-b0e9a032a433\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.208511 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/597ad435-7283-469d-8c9a-b0e9a032a433-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm\" (UID: \"597ad435-7283-469d-8c9a-b0e9a032a433\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.212369 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw\" (UID: \"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.226495 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lk47\" (UniqueName: \"kubernetes.io/projected/ad4b53c3-5d16-412b-ba96-f4ba1a54624a-kube-api-access-2lk47\") pod \"obo-prometheus-operator-668cf9dfbb-jvrtq\" (UID: \"ad4b53c3-5d16-412b-ba96-f4ba1a54624a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.228845 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-7mc44"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.284555 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.287200 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzzp5\" (UniqueName: \"kubernetes.io/projected/9bd8cd77-0f23-40cc-87aa-19ef553d9565-kube-api-access-fzzp5\") pod \"observability-operator-d8bb48f5d-7mc44\" (UID: \"9bd8cd77-0f23-40cc-87aa-19ef553d9565\") " pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.287432 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bd8cd77-0f23-40cc-87aa-19ef553d9565-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-7mc44\" (UID: \"9bd8cd77-0f23-40cc-87aa-19ef553d9565\") " pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.380119 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.388885 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzzp5\" (UniqueName: \"kubernetes.io/projected/9bd8cd77-0f23-40cc-87aa-19ef553d9565-kube-api-access-fzzp5\") pod \"observability-operator-d8bb48f5d-7mc44\" (UID: \"9bd8cd77-0f23-40cc-87aa-19ef553d9565\") " pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.388990 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bd8cd77-0f23-40cc-87aa-19ef553d9565-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-7mc44\" (UID: \"9bd8cd77-0f23-40cc-87aa-19ef553d9565\") " pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.390145 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.395244 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9bd8cd77-0f23-40cc-87aa-19ef553d9565-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-7mc44\" (UID: \"9bd8cd77-0f23-40cc-87aa-19ef553d9565\") " pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.410614 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-gcb2c"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.412427 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.418514 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzzp5\" (UniqueName: \"kubernetes.io/projected/9bd8cd77-0f23-40cc-87aa-19ef553d9565-kube-api-access-fzzp5\") pod \"observability-operator-d8bb48f5d-7mc44\" (UID: \"9bd8cd77-0f23-40cc-87aa-19ef553d9565\") " pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.425210 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-r42k4" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.426347 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-gcb2c"] Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.494848 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f-openshift-service-ca\") pod \"perses-operator-5446b9c989-gcb2c\" (UID: \"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f\") " pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.495266 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlwnr\" (UniqueName: \"kubernetes.io/projected/334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f-kube-api-access-tlwnr\") pod \"perses-operator-5446b9c989-gcb2c\" (UID: \"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f\") " pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.600035 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f-openshift-service-ca\") pod \"perses-operator-5446b9c989-gcb2c\" (UID: \"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f\") " pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.600154 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlwnr\" (UniqueName: \"kubernetes.io/projected/334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f-kube-api-access-tlwnr\") pod \"perses-operator-5446b9c989-gcb2c\" (UID: \"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f\") " pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.602296 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f-openshift-service-ca\") pod \"perses-operator-5446b9c989-gcb2c\" (UID: \"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f\") " pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.620516 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlwnr\" (UniqueName: \"kubernetes.io/projected/334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f-kube-api-access-tlwnr\") pod \"perses-operator-5446b9c989-gcb2c\" (UID: \"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f\") " pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.625462 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.809113 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:21:29 crc kubenswrapper[4817]: I1128 15:21:29.954536 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq"] Nov 28 15:21:29 crc kubenswrapper[4817]: W1128 15:21:29.973127 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad4b53c3_5d16_412b_ba96_f4ba1a54624a.slice/crio-f0e72b2f6b2fd48fca0fe525db6930dab459d96b883dbc1431f401a0b8e9f3bb WatchSource:0}: Error finding container f0e72b2f6b2fd48fca0fe525db6930dab459d96b883dbc1431f401a0b8e9f3bb: Status 404 returned error can't find the container with id f0e72b2f6b2fd48fca0fe525db6930dab459d96b883dbc1431f401a0b8e9f3bb Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.035830 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw"] Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.063969 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm"] Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.303178 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-7mc44"] Nov 28 15:21:30 crc kubenswrapper[4817]: W1128 15:21:30.332832 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bd8cd77_0f23_40cc_87aa_19ef553d9565.slice/crio-3fae83120168527879e587599acc2c75ebe0a29f81f8a65561c8658f9cb196bb WatchSource:0}: Error finding container 3fae83120168527879e587599acc2c75ebe0a29f81f8a65561c8658f9cb196bb: Status 404 returned error can't find the container with id 3fae83120168527879e587599acc2c75ebe0a29f81f8a65561c8658f9cb196bb Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.433274 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-gcb2c"] Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.557483 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" event={"ID":"9bd8cd77-0f23-40cc-87aa-19ef553d9565","Type":"ContainerStarted","Data":"3fae83120168527879e587599acc2c75ebe0a29f81f8a65561c8658f9cb196bb"} Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.559015 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" event={"ID":"ad4b53c3-5d16-412b-ba96-f4ba1a54624a","Type":"ContainerStarted","Data":"f0e72b2f6b2fd48fca0fe525db6930dab459d96b883dbc1431f401a0b8e9f3bb"} Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.560301 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" event={"ID":"597ad435-7283-469d-8c9a-b0e9a032a433","Type":"ContainerStarted","Data":"f1c869559880b4f720e74736b1c0014ad875cf314664992a75b60477db705a1f"} Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.561317 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" event={"ID":"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf","Type":"ContainerStarted","Data":"b45239b708ca8cdb56d1d34291afa93c7c3c3012941c5718de501651303fec4e"} Nov 28 15:21:30 crc kubenswrapper[4817]: I1128 15:21:30.562409 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" event={"ID":"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f","Type":"ContainerStarted","Data":"3d5fe564dc1d18bb8f395e3636f5c53459eb551fd1351d16448c32d28d44b758"} Nov 28 15:21:37 crc kubenswrapper[4817]: I1128 15:21:37.751805 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:21:37 crc kubenswrapper[4817]: E1128 15:21:37.755315 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:21:46 crc kubenswrapper[4817]: E1128 15:21:46.967509 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Nov 28 15:21:46 crc kubenswrapper[4817]: E1128 15:21:46.968084 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fzzp5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-7mc44_openshift-operators(9bd8cd77-0f23-40cc-87aa-19ef553d9565): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:21:46 crc kubenswrapper[4817]: E1128 15:21:46.969426 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" podUID="9bd8cd77-0f23-40cc-87aa-19ef553d9565" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.005387 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.005521 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm_openshift-operators(597ad435-7283-469d-8c9a-b0e9a032a433): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.006835 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" podUID="597ad435-7283-469d-8c9a-b0e9a032a433" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.531171 4817 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.531330 4817 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tlwnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-gcb2c_openshift-operators(334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.532517 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" podUID="334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.747281 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" podUID="9bd8cd77-0f23-40cc-87aa-19ef553d9565" Nov 28 15:21:47 crc kubenswrapper[4817]: E1128 15:21:47.747302 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" podUID="334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f" Nov 28 15:21:48 crc kubenswrapper[4817]: I1128 15:21:48.753708 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" event={"ID":"ad4b53c3-5d16-412b-ba96-f4ba1a54624a","Type":"ContainerStarted","Data":"47bb07e201c5ec46790ec55e3856e9b6af037a1591dec40e5f9b749a8b2a01f0"} Nov 28 15:21:48 crc kubenswrapper[4817]: I1128 15:21:48.757310 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" event={"ID":"597ad435-7283-469d-8c9a-b0e9a032a433","Type":"ContainerStarted","Data":"358fa1ef9a6ef3db5053bfb2b0456e73c3a60a4087819b28d4608024b2b041e2"} Nov 28 15:21:48 crc kubenswrapper[4817]: I1128 15:21:48.765264 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" event={"ID":"4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf","Type":"ContainerStarted","Data":"4af35c186aff697a11590558684c734775d78fbf3ce0517d84cfe7f7c8a88880"} Nov 28 15:21:48 crc kubenswrapper[4817]: I1128 15:21:48.783644 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-jvrtq" podStartSLOduration=3.234898414 podStartE2EDuration="20.783616157s" podCreationTimestamp="2025-11-28 15:21:28 +0000 UTC" firstStartedPulling="2025-11-28 15:21:29.977929654 +0000 UTC m=+3112.565907920" lastFinishedPulling="2025-11-28 15:21:47.526647397 +0000 UTC m=+3130.114625663" observedRunningTime="2025-11-28 15:21:48.770543705 +0000 UTC m=+3131.358521971" watchObservedRunningTime="2025-11-28 15:21:48.783616157 +0000 UTC m=+3131.371594423" Nov 28 15:21:48 crc kubenswrapper[4817]: I1128 15:21:48.811000 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw" podStartSLOduration=2.327619147 podStartE2EDuration="19.810979471s" podCreationTimestamp="2025-11-28 15:21:29 +0000 UTC" firstStartedPulling="2025-11-28 15:21:30.045691263 +0000 UTC m=+3112.633669529" lastFinishedPulling="2025-11-28 15:21:47.529051577 +0000 UTC m=+3130.117029853" observedRunningTime="2025-11-28 15:21:48.807123793 +0000 UTC m=+3131.395102079" watchObservedRunningTime="2025-11-28 15:21:48.810979471 +0000 UTC m=+3131.398957737" Nov 28 15:21:48 crc kubenswrapper[4817]: I1128 15:21:48.880240 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm" podStartSLOduration=-9223372015.974567 podStartE2EDuration="20.880208167s" podCreationTimestamp="2025-11-28 15:21:28 +0000 UTC" firstStartedPulling="2025-11-28 15:21:30.094005629 +0000 UTC m=+3112.681983895" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:21:48.842263775 +0000 UTC m=+3131.430242041" watchObservedRunningTime="2025-11-28 15:21:48.880208167 +0000 UTC m=+3131.468186433" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.075416 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rc2bq"] Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.078062 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.101206 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rc2bq"] Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.187589 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-catalog-content\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.187868 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7fl9\" (UniqueName: \"kubernetes.io/projected/0d92b696-0bac-4e37-9c7e-7785ca9d199b-kube-api-access-m7fl9\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.188010 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-utilities\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.290193 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-catalog-content\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.290280 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7fl9\" (UniqueName: \"kubernetes.io/projected/0d92b696-0bac-4e37-9c7e-7785ca9d199b-kube-api-access-m7fl9\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.290367 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-utilities\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.290781 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-catalog-content\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.290865 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-utilities\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.312516 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7fl9\" (UniqueName: \"kubernetes.io/projected/0d92b696-0bac-4e37-9c7e-7785ca9d199b-kube-api-access-m7fl9\") pod \"community-operators-rc2bq\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.395790 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.762114 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rc2bq"] Nov 28 15:21:50 crc kubenswrapper[4817]: W1128 15:21:50.775676 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d92b696_0bac_4e37_9c7e_7785ca9d199b.slice/crio-97c0468cecf731ae1c1b151831be3a122651880e8d53155a90ade1be76a9405f WatchSource:0}: Error finding container 97c0468cecf731ae1c1b151831be3a122651880e8d53155a90ade1be76a9405f: Status 404 returned error can't find the container with id 97c0468cecf731ae1c1b151831be3a122651880e8d53155a90ade1be76a9405f Nov 28 15:21:50 crc kubenswrapper[4817]: I1128 15:21:50.806809 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerStarted","Data":"97c0468cecf731ae1c1b151831be3a122651880e8d53155a90ade1be76a9405f"} Nov 28 15:21:51 crc kubenswrapper[4817]: I1128 15:21:51.737532 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:21:51 crc kubenswrapper[4817]: E1128 15:21:51.738139 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:21:51 crc kubenswrapper[4817]: I1128 15:21:51.818351 4817 generic.go:334] "Generic (PLEG): container finished" podID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerID="01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352" exitCode=0 Nov 28 15:21:51 crc kubenswrapper[4817]: I1128 15:21:51.818411 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerDied","Data":"01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352"} Nov 28 15:21:52 crc kubenswrapper[4817]: I1128 15:21:52.828234 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerStarted","Data":"cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718"} Nov 28 15:21:53 crc kubenswrapper[4817]: I1128 15:21:53.838680 4817 generic.go:334] "Generic (PLEG): container finished" podID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerID="cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718" exitCode=0 Nov 28 15:21:53 crc kubenswrapper[4817]: I1128 15:21:53.838759 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerDied","Data":"cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718"} Nov 28 15:21:54 crc kubenswrapper[4817]: I1128 15:21:54.849327 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerStarted","Data":"524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2"} Nov 28 15:21:54 crc kubenswrapper[4817]: I1128 15:21:54.874556 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rc2bq" podStartSLOduration=2.484064602 podStartE2EDuration="4.874537852s" podCreationTimestamp="2025-11-28 15:21:50 +0000 UTC" firstStartedPulling="2025-11-28 15:21:51.821930872 +0000 UTC m=+3134.409909158" lastFinishedPulling="2025-11-28 15:21:54.212404142 +0000 UTC m=+3136.800382408" observedRunningTime="2025-11-28 15:21:54.86932671 +0000 UTC m=+3137.457304976" watchObservedRunningTime="2025-11-28 15:21:54.874537852 +0000 UTC m=+3137.462516118" Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.065007 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.066244 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-api" containerID="cri-o://e6342f7d19900580ceafb5ceea16f4586d9af2ea0ebb4304e8d091d131d2d19d" gracePeriod=30 Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.066618 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-listener" containerID="cri-o://f00192d3b73dfcd6b0e6da9485fa02b33d0386b0c9ae8f16ac6e3384a6409b16" gracePeriod=30 Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.066656 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-notifier" containerID="cri-o://8052cf05e1ff2893d30364ce1fd7349dd3c60eccd6e24f7d9ce87cc1249c3bd9" gracePeriod=30 Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.066696 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-evaluator" containerID="cri-o://2f656a9de3fcca471f43f632b6e9fb00a7e1676c2d800c633566136f16e79d8a" gracePeriod=30 Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.877907 4817 generic.go:334] "Generic (PLEG): container finished" podID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerID="2f656a9de3fcca471f43f632b6e9fb00a7e1676c2d800c633566136f16e79d8a" exitCode=0 Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.878338 4817 generic.go:334] "Generic (PLEG): container finished" podID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerID="e6342f7d19900580ceafb5ceea16f4586d9af2ea0ebb4304e8d091d131d2d19d" exitCode=0 Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.877985 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerDied","Data":"2f656a9de3fcca471f43f632b6e9fb00a7e1676c2d800c633566136f16e79d8a"} Nov 28 15:21:56 crc kubenswrapper[4817]: I1128 15:21:56.878477 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerDied","Data":"e6342f7d19900580ceafb5ceea16f4586d9af2ea0ebb4304e8d091d131d2d19d"} Nov 28 15:21:59 crc kubenswrapper[4817]: I1128 15:21:59.916646 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" event={"ID":"9bd8cd77-0f23-40cc-87aa-19ef553d9565","Type":"ContainerStarted","Data":"3e3fac2af67e4424e6fd44d41ee4fae3561aba0d0298bc13600055fbe68e20db"} Nov 28 15:21:59 crc kubenswrapper[4817]: I1128 15:21:59.917841 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:21:59 crc kubenswrapper[4817]: I1128 15:21:59.919319 4817 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-7mc44 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.6:8081/healthz\": dial tcp 10.217.1.6:8081: connect: connection refused" start-of-body= Nov 28 15:21:59 crc kubenswrapper[4817]: I1128 15:21:59.919379 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" podUID="9bd8cd77-0f23-40cc-87aa-19ef553d9565" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.6:8081/healthz\": dial tcp 10.217.1.6:8081: connect: connection refused" Nov 28 15:21:59 crc kubenswrapper[4817]: I1128 15:21:59.952524 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" podStartSLOduration=1.749351496 podStartE2EDuration="30.952506926s" podCreationTimestamp="2025-11-28 15:21:29 +0000 UTC" firstStartedPulling="2025-11-28 15:21:30.365923617 +0000 UTC m=+3112.953901883" lastFinishedPulling="2025-11-28 15:21:59.569079047 +0000 UTC m=+3142.157057313" observedRunningTime="2025-11-28 15:21:59.940952262 +0000 UTC m=+3142.528930528" watchObservedRunningTime="2025-11-28 15:21:59.952506926 +0000 UTC m=+3142.540485192" Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.396063 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.396966 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.495119 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.927301 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" event={"ID":"334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f","Type":"ContainerStarted","Data":"c1daa43922aed6bddd74318975a137309c6649f07b09f922ca2beb67a2fec7ad"} Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.927874 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.935439 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-7mc44" Nov 28 15:22:00 crc kubenswrapper[4817]: I1128 15:22:00.979401 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" podStartSLOduration=1.8646407900000002 podStartE2EDuration="31.979376259s" podCreationTimestamp="2025-11-28 15:21:29 +0000 UTC" firstStartedPulling="2025-11-28 15:21:30.448757828 +0000 UTC m=+3113.036736094" lastFinishedPulling="2025-11-28 15:22:00.563493277 +0000 UTC m=+3143.151471563" observedRunningTime="2025-11-28 15:22:00.959660819 +0000 UTC m=+3143.547639095" watchObservedRunningTime="2025-11-28 15:22:00.979376259 +0000 UTC m=+3143.567354535" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.091593 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.207267 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rc2bq"] Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.813480 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.816250 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.818808 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.818907 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.819177 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.819426 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.819564 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-8fljr" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831058 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9f575d2e-8149-4fad-86a3-da8552e0e9e0-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831165 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831223 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831261 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831297 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9f575d2e-8149-4fad-86a3-da8552e0e9e0-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831357 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnzx4\" (UniqueName: \"kubernetes.io/projected/9f575d2e-8149-4fad-86a3-da8552e0e9e0-kube-api-access-jnzx4\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.831453 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9f575d2e-8149-4fad-86a3-da8552e0e9e0-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.846697 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933173 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9f575d2e-8149-4fad-86a3-da8552e0e9e0-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933296 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9f575d2e-8149-4fad-86a3-da8552e0e9e0-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933333 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933356 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933389 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933437 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9f575d2e-8149-4fad-86a3-da8552e0e9e0-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.933469 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnzx4\" (UniqueName: \"kubernetes.io/projected/9f575d2e-8149-4fad-86a3-da8552e0e9e0-kube-api-access-jnzx4\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.935906 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/9f575d2e-8149-4fad-86a3-da8552e0e9e0-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.939437 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.939437 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.940215 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9f575d2e-8149-4fad-86a3-da8552e0e9e0-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.942817 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9f575d2e-8149-4fad-86a3-da8552e0e9e0-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.943038 4817 generic.go:334] "Generic (PLEG): container finished" podID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerID="f00192d3b73dfcd6b0e6da9485fa02b33d0386b0c9ae8f16ac6e3384a6409b16" exitCode=0 Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.943922 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerDied","Data":"f00192d3b73dfcd6b0e6da9485fa02b33d0386b0c9ae8f16ac6e3384a6409b16"} Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.947281 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/9f575d2e-8149-4fad-86a3-da8552e0e9e0-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:01 crc kubenswrapper[4817]: I1128 15:22:01.962430 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnzx4\" (UniqueName: \"kubernetes.io/projected/9f575d2e-8149-4fad-86a3-da8552e0e9e0-kube-api-access-jnzx4\") pod \"alertmanager-metric-storage-0\" (UID: \"9f575d2e-8149-4fad-86a3-da8552e0e9e0\") " pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.134072 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.423168 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.425991 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.430437 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.430586 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.430669 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.430751 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.430878 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.430943 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jkmqg" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.433630 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.441802 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ff3653e-1322-4868-89d5-13179cd71578-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.441859 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.441898 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ff3653e-1322-4868-89d5-13179cd71578-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.441931 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.441958 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.442021 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dhxc\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-kube-api-access-4dhxc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.442042 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.442067 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.544272 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.544533 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dhxc\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-kube-api-access-4dhxc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.544609 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.544697 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.544854 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ff3653e-1322-4868-89d5-13179cd71578-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.544980 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.545085 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ff3653e-1322-4868-89d5-13179cd71578-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.545181 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.545297 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.546556 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ff3653e-1322-4868-89d5-13179cd71578-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.549310 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ff3653e-1322-4868-89d5-13179cd71578-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.551314 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.551619 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.551861 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.552673 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.560856 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dhxc\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-kube-api-access-4dhxc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.594218 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.632221 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 28 15:22:02 crc kubenswrapper[4817]: W1128 15:22:02.637576 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f575d2e_8149_4fad_86a3_da8552e0e9e0.slice/crio-cbb3b79f764d0d9fd880d6f58a98536bcd75867020ff932fd5a0e6d8140e1d8d WatchSource:0}: Error finding container cbb3b79f764d0d9fd880d6f58a98536bcd75867020ff932fd5a0e6d8140e1d8d: Status 404 returned error can't find the container with id cbb3b79f764d0d9fd880d6f58a98536bcd75867020ff932fd5a0e6d8140e1d8d Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.739999 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:22:02 crc kubenswrapper[4817]: E1128 15:22:02.740840 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.752614 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.964553 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"9f575d2e-8149-4fad-86a3-da8552e0e9e0","Type":"ContainerStarted","Data":"cbb3b79f764d0d9fd880d6f58a98536bcd75867020ff932fd5a0e6d8140e1d8d"} Nov 28 15:22:02 crc kubenswrapper[4817]: I1128 15:22:02.964739 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rc2bq" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="registry-server" containerID="cri-o://524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2" gracePeriod=2 Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.057324 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.721649 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.873275 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-utilities\") pod \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.873567 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-catalog-content\") pod \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.873662 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7fl9\" (UniqueName: \"kubernetes.io/projected/0d92b696-0bac-4e37-9c7e-7785ca9d199b-kube-api-access-m7fl9\") pod \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\" (UID: \"0d92b696-0bac-4e37-9c7e-7785ca9d199b\") " Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.875280 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-utilities" (OuterVolumeSpecName: "utilities") pod "0d92b696-0bac-4e37-9c7e-7785ca9d199b" (UID: "0d92b696-0bac-4e37-9c7e-7785ca9d199b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.880970 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d92b696-0bac-4e37-9c7e-7785ca9d199b-kube-api-access-m7fl9" (OuterVolumeSpecName: "kube-api-access-m7fl9") pod "0d92b696-0bac-4e37-9c7e-7785ca9d199b" (UID: "0d92b696-0bac-4e37-9c7e-7785ca9d199b"). InnerVolumeSpecName "kube-api-access-m7fl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.928771 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d92b696-0bac-4e37-9c7e-7785ca9d199b" (UID: "0d92b696-0bac-4e37-9c7e-7785ca9d199b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.974867 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerStarted","Data":"68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a"} Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.975945 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.975971 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d92b696-0bac-4e37-9c7e-7785ca9d199b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.975984 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7fl9\" (UniqueName: \"kubernetes.io/projected/0d92b696-0bac-4e37-9c7e-7785ca9d199b-kube-api-access-m7fl9\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.977639 4817 generic.go:334] "Generic (PLEG): container finished" podID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerID="524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2" exitCode=0 Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.977666 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerDied","Data":"524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2"} Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.977708 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rc2bq" event={"ID":"0d92b696-0bac-4e37-9c7e-7785ca9d199b","Type":"ContainerDied","Data":"97c0468cecf731ae1c1b151831be3a122651880e8d53155a90ade1be76a9405f"} Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.977750 4817 scope.go:117] "RemoveContainer" containerID="524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2" Nov 28 15:22:03 crc kubenswrapper[4817]: I1128 15:22:03.977771 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rc2bq" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.025557 4817 scope.go:117] "RemoveContainer" containerID="cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.063864 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rc2bq"] Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.071237 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rc2bq"] Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.093660 4817 scope.go:117] "RemoveContainer" containerID="01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.123105 4817 scope.go:117] "RemoveContainer" containerID="524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2" Nov 28 15:22:04 crc kubenswrapper[4817]: E1128 15:22:04.123654 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2\": container with ID starting with 524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2 not found: ID does not exist" containerID="524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.123701 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2"} err="failed to get container status \"524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2\": rpc error: code = NotFound desc = could not find container \"524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2\": container with ID starting with 524b15f4a17ce2031992965f26858409c2a734999900357ac372c1fc8b41b1f2 not found: ID does not exist" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.123747 4817 scope.go:117] "RemoveContainer" containerID="cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718" Nov 28 15:22:04 crc kubenswrapper[4817]: E1128 15:22:04.124136 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718\": container with ID starting with cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718 not found: ID does not exist" containerID="cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.124254 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718"} err="failed to get container status \"cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718\": rpc error: code = NotFound desc = could not find container \"cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718\": container with ID starting with cb71f43ad68741d137f08fcdb48ae17d338922b4f73b4c6d5d0f9c88d2083718 not found: ID does not exist" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.124300 4817 scope.go:117] "RemoveContainer" containerID="01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352" Nov 28 15:22:04 crc kubenswrapper[4817]: E1128 15:22:04.124660 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352\": container with ID starting with 01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352 not found: ID does not exist" containerID="01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352" Nov 28 15:22:04 crc kubenswrapper[4817]: I1128 15:22:04.124778 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352"} err="failed to get container status \"01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352\": rpc error: code = NotFound desc = could not find container \"01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352\": container with ID starting with 01872d38db6416daa51d50921dbc8e0a963e1ffcd8ba42ea6c52b8c11532b352 not found: ID does not exist" Nov 28 15:22:05 crc kubenswrapper[4817]: I1128 15:22:05.748572 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" path="/var/lib/kubelet/pods/0d92b696-0bac-4e37-9c7e-7785ca9d199b/volumes" Nov 28 15:22:09 crc kubenswrapper[4817]: I1128 15:22:09.037514 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerStarted","Data":"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9"} Nov 28 15:22:09 crc kubenswrapper[4817]: I1128 15:22:09.811835 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-gcb2c" Nov 28 15:22:10 crc kubenswrapper[4817]: I1128 15:22:10.046956 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"9f575d2e-8149-4fad-86a3-da8552e0e9e0","Type":"ContainerStarted","Data":"9e112d4c1f788a376378476cd7e3de67501baf356619e2e0388df3498f14b876"} Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.090237 4817 generic.go:334] "Generic (PLEG): container finished" podID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerID="8052cf05e1ff2893d30364ce1fd7349dd3c60eccd6e24f7d9ce87cc1249c3bd9" exitCode=0 Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.091331 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerDied","Data":"8052cf05e1ff2893d30364ce1fd7349dd3c60eccd6e24f7d9ce87cc1249c3bd9"} Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.091356 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c","Type":"ContainerDied","Data":"caf819c0569f239e71e207bbaeaefa0bcc5d2cc31224ade3ca24af9cb1d19f02"} Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.091365 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caf819c0569f239e71e207bbaeaefa0bcc5d2cc31224ade3ca24af9cb1d19f02" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.126393 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.225011 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-public-tls-certs\") pod \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.225054 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-combined-ca-bundle\") pod \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.225078 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfv9x\" (UniqueName: \"kubernetes.io/projected/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-kube-api-access-pfv9x\") pod \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.225111 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-internal-tls-certs\") pod \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.225230 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-config-data\") pod \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.225278 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-scripts\") pod \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\" (UID: \"332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c\") " Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.247829 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-scripts" (OuterVolumeSpecName: "scripts") pod "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" (UID: "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.247951 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-kube-api-access-pfv9x" (OuterVolumeSpecName: "kube-api-access-pfv9x") pod "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" (UID: "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c"). InnerVolumeSpecName "kube-api-access-pfv9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.328019 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.328160 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfv9x\" (UniqueName: \"kubernetes.io/projected/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-kube-api-access-pfv9x\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.342888 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" (UID: "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.400818 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" (UID: "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.431792 4817 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.431828 4817 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.448918 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-config-data" (OuterVolumeSpecName: "config-data") pod "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" (UID: "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.451806 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" (UID: "332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.533906 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:11 crc kubenswrapper[4817]: I1128 15:22:11.533939 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.098173 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.123087 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.131335 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.156884 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.157536 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="extract-utilities" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.157667 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="extract-utilities" Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.157820 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-api" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.157909 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-api" Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.157995 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-listener" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.158078 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-listener" Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.158166 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="registry-server" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.158246 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="registry-server" Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.158325 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-notifier" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.158398 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-notifier" Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.158464 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="extract-content" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.158537 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="extract-content" Nov 28 15:22:12 crc kubenswrapper[4817]: E1128 15:22:12.158630 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-evaluator" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.158696 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-evaluator" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.159026 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-listener" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.159142 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-evaluator" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.159236 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d92b696-0bac-4e37-9c7e-7785ca9d199b" containerName="registry-server" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.159347 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-api" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.159427 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" containerName="aodh-notifier" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.165802 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.172258 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.174004 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.174429 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.174582 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dzj58" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.174696 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.187857 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.245743 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-internal-tls-certs\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.245794 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-combined-ca-bundle\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.245821 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z57fl\" (UniqueName: \"kubernetes.io/projected/471e9faa-e3ae-4f2c-8894-d6a84ec97417-kube-api-access-z57fl\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.245892 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-config-data\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.245921 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-scripts\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.245976 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-public-tls-certs\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.348100 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-internal-tls-certs\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.348148 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-combined-ca-bundle\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.348172 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z57fl\" (UniqueName: \"kubernetes.io/projected/471e9faa-e3ae-4f2c-8894-d6a84ec97417-kube-api-access-z57fl\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.348209 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-config-data\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.348234 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-scripts\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.348285 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-public-tls-certs\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.356378 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-scripts\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.356735 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-internal-tls-certs\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.359285 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-combined-ca-bundle\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.360937 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-config-data\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.378183 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-public-tls-certs\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.385558 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z57fl\" (UniqueName: \"kubernetes.io/projected/471e9faa-e3ae-4f2c-8894-d6a84ec97417-kube-api-access-z57fl\") pod \"aodh-0\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " pod="openstack/aodh-0" Nov 28 15:22:12 crc kubenswrapper[4817]: I1128 15:22:12.543882 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:13 crc kubenswrapper[4817]: I1128 15:22:13.075746 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:13 crc kubenswrapper[4817]: I1128 15:22:13.082629 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:22:13 crc kubenswrapper[4817]: I1128 15:22:13.107606 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerStarted","Data":"7502c793d0bcbd74aea5fc85d77d60e18b6c2caf104dcac0d9a9b26509006bc2"} Nov 28 15:22:13 crc kubenswrapper[4817]: I1128 15:22:13.737623 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:22:13 crc kubenswrapper[4817]: E1128 15:22:13.738158 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:22:13 crc kubenswrapper[4817]: I1128 15:22:13.747403 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c" path="/var/lib/kubelet/pods/332f7b9f-dd4c-4bde-b1d2-3bbb06f8701c/volumes" Nov 28 15:22:15 crc kubenswrapper[4817]: I1128 15:22:15.126659 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ff3653e-1322-4868-89d5-13179cd71578" containerID="a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9" exitCode=0 Nov 28 15:22:15 crc kubenswrapper[4817]: I1128 15:22:15.126763 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerDied","Data":"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9"} Nov 28 15:22:15 crc kubenswrapper[4817]: I1128 15:22:15.128628 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerStarted","Data":"491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d"} Nov 28 15:22:15 crc kubenswrapper[4817]: I1128 15:22:15.130644 4817 generic.go:334] "Generic (PLEG): container finished" podID="9f575d2e-8149-4fad-86a3-da8552e0e9e0" containerID="9e112d4c1f788a376378476cd7e3de67501baf356619e2e0388df3498f14b876" exitCode=0 Nov 28 15:22:15 crc kubenswrapper[4817]: I1128 15:22:15.130694 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"9f575d2e-8149-4fad-86a3-da8552e0e9e0","Type":"ContainerDied","Data":"9e112d4c1f788a376378476cd7e3de67501baf356619e2e0388df3498f14b876"} Nov 28 15:22:16 crc kubenswrapper[4817]: I1128 15:22:16.145075 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerStarted","Data":"9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76"} Nov 28 15:22:16 crc kubenswrapper[4817]: I1128 15:22:16.146935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerStarted","Data":"a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643"} Nov 28 15:22:17 crc kubenswrapper[4817]: I1128 15:22:17.156837 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerStarted","Data":"ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5"} Nov 28 15:22:17 crc kubenswrapper[4817]: I1128 15:22:17.185501 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.682213915 podStartE2EDuration="5.185479467s" podCreationTimestamp="2025-11-28 15:22:12 +0000 UTC" firstStartedPulling="2025-11-28 15:22:13.082446099 +0000 UTC m=+3155.670424365" lastFinishedPulling="2025-11-28 15:22:16.585711651 +0000 UTC m=+3159.173689917" observedRunningTime="2025-11-28 15:22:17.178147181 +0000 UTC m=+3159.766125457" watchObservedRunningTime="2025-11-28 15:22:17.185479467 +0000 UTC m=+3159.773457733" Nov 28 15:22:24 crc kubenswrapper[4817]: I1128 15:22:24.238697 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"9f575d2e-8149-4fad-86a3-da8552e0e9e0","Type":"ContainerStarted","Data":"98164a12a6d68b5acf5c2a277e4b25c2c5f92554ae0c7bbc2e0223d3557dadbe"} Nov 28 15:22:24 crc kubenswrapper[4817]: I1128 15:22:24.240956 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerStarted","Data":"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da"} Nov 28 15:22:27 crc kubenswrapper[4817]: I1128 15:22:27.746593 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:22:27 crc kubenswrapper[4817]: E1128 15:22:27.747450 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:22:28 crc kubenswrapper[4817]: I1128 15:22:28.291663 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"9f575d2e-8149-4fad-86a3-da8552e0e9e0","Type":"ContainerStarted","Data":"b32a60778eba4207b383b005451d08044fd1edfa62a9a47d6105f97d9f83ad8f"} Nov 28 15:22:28 crc kubenswrapper[4817]: I1128 15:22:28.293108 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:28 crc kubenswrapper[4817]: I1128 15:22:28.296546 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 28 15:22:28 crc kubenswrapper[4817]: I1128 15:22:28.322508 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.4277896420000005 podStartE2EDuration="27.322475797s" podCreationTimestamp="2025-11-28 15:22:01 +0000 UTC" firstStartedPulling="2025-11-28 15:22:02.643087769 +0000 UTC m=+3145.231066035" lastFinishedPulling="2025-11-28 15:22:23.537773904 +0000 UTC m=+3166.125752190" observedRunningTime="2025-11-28 15:22:28.315946791 +0000 UTC m=+3170.903925077" watchObservedRunningTime="2025-11-28 15:22:28.322475797 +0000 UTC m=+3170.910454103" Nov 28 15:22:29 crc kubenswrapper[4817]: I1128 15:22:29.303679 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerStarted","Data":"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c"} Nov 28 15:22:32 crc kubenswrapper[4817]: I1128 15:22:32.332700 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerStarted","Data":"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55"} Nov 28 15:22:32 crc kubenswrapper[4817]: I1128 15:22:32.358220 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=2.983513027 podStartE2EDuration="31.358177678s" podCreationTimestamp="2025-11-28 15:22:01 +0000 UTC" firstStartedPulling="2025-11-28 15:22:03.061691659 +0000 UTC m=+3145.649669915" lastFinishedPulling="2025-11-28 15:22:31.4363563 +0000 UTC m=+3174.024334566" observedRunningTime="2025-11-28 15:22:32.353988641 +0000 UTC m=+3174.941966907" watchObservedRunningTime="2025-11-28 15:22:32.358177678 +0000 UTC m=+3174.946155944" Nov 28 15:22:32 crc kubenswrapper[4817]: I1128 15:22:32.753252 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:32 crc kubenswrapper[4817]: I1128 15:22:32.753310 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:32 crc kubenswrapper[4817]: I1128 15:22:32.755938 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:33 crc kubenswrapper[4817]: I1128 15:22:33.348115 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.673185 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.673684 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" containerName="openstackclient" containerID="cri-o://1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b" gracePeriod=2 Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.687090 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.697043 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 15:22:34 crc kubenswrapper[4817]: E1128 15:22:34.697439 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" containerName="openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.697456 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" containerName="openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.697647 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" containerName="openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.698335 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.711876 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.715893 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" podUID="6df83271-9eeb-4eee-b838-feec85dde399" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.806509 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6df83271-9eeb-4eee-b838-feec85dde399-openstack-config-secret\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.806598 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df83271-9eeb-4eee-b838-feec85dde399-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.806647 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6df83271-9eeb-4eee-b838-feec85dde399-openstack-config\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.806758 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq4x9\" (UniqueName: \"kubernetes.io/projected/6df83271-9eeb-4eee-b838-feec85dde399-kube-api-access-sq4x9\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.908686 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6df83271-9eeb-4eee-b838-feec85dde399-openstack-config-secret\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.908789 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df83271-9eeb-4eee-b838-feec85dde399-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.908837 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6df83271-9eeb-4eee-b838-feec85dde399-openstack-config\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.908876 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq4x9\" (UniqueName: \"kubernetes.io/projected/6df83271-9eeb-4eee-b838-feec85dde399-kube-api-access-sq4x9\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.909854 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6df83271-9eeb-4eee-b838-feec85dde399-openstack-config\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.913960 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6df83271-9eeb-4eee-b838-feec85dde399-openstack-config-secret\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.924212 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df83271-9eeb-4eee-b838-feec85dde399-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.924800 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq4x9\" (UniqueName: \"kubernetes.io/projected/6df83271-9eeb-4eee-b838-feec85dde399-kube-api-access-sq4x9\") pod \"openstackclient\" (UID: \"6df83271-9eeb-4eee-b838-feec85dde399\") " pod="openstack/openstackclient" Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.991851 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.992454 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-api" containerID="cri-o://491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d" gracePeriod=30 Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.992667 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-listener" containerID="cri-o://ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5" gracePeriod=30 Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.992779 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-evaluator" containerID="cri-o://a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643" gracePeriod=30 Nov 28 15:22:34 crc kubenswrapper[4817]: I1128 15:22:34.992953 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-notifier" containerID="cri-o://9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76" gracePeriod=30 Nov 28 15:22:35 crc kubenswrapper[4817]: I1128 15:22:35.018512 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:35.440147 4817 generic.go:334] "Generic (PLEG): container finished" podID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerID="491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d" exitCode=0 Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:35.440809 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerDied","Data":"491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d"} Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:35.888344 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.312676 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 15:22:36 crc kubenswrapper[4817]: W1128 15:22:36.314944 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df83271_9eeb_4eee_b838_feec85dde399.slice/crio-b61009f405fd1dbdf11957b6f9d171f92d02da6fa9a7684e444a11786ce17433 WatchSource:0}: Error finding container b61009f405fd1dbdf11957b6f9d171f92d02da6fa9a7684e444a11786ce17433: Status 404 returned error can't find the container with id b61009f405fd1dbdf11957b6f9d171f92d02da6fa9a7684e444a11786ce17433 Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.457458 4817 generic.go:334] "Generic (PLEG): container finished" podID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerID="a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643" exitCode=0 Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.457536 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerDied","Data":"a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643"} Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.459494 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6df83271-9eeb-4eee-b838-feec85dde399","Type":"ContainerStarted","Data":"b61009f405fd1dbdf11957b6f9d171f92d02da6fa9a7684e444a11786ce17433"} Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.459678 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="prometheus" containerID="cri-o://80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" gracePeriod=600 Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.460077 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="thanos-sidecar" containerID="cri-o://f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" gracePeriod=600 Nov 28 15:22:36 crc kubenswrapper[4817]: I1128 15:22:36.460133 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="config-reloader" containerID="cri-o://822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" gracePeriod=600 Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.113214 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.255229 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config\") pod \"efe3c20f-1ca9-46ef-92cd-de439ba78250\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.255441 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmc2g\" (UniqueName: \"kubernetes.io/projected/efe3c20f-1ca9-46ef-92cd-de439ba78250-kube-api-access-dmc2g\") pod \"efe3c20f-1ca9-46ef-92cd-de439ba78250\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.255517 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config-secret\") pod \"efe3c20f-1ca9-46ef-92cd-de439ba78250\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.255545 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-combined-ca-bundle\") pod \"efe3c20f-1ca9-46ef-92cd-de439ba78250\" (UID: \"efe3c20f-1ca9-46ef-92cd-de439ba78250\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.260886 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe3c20f-1ca9-46ef-92cd-de439ba78250-kube-api-access-dmc2g" (OuterVolumeSpecName: "kube-api-access-dmc2g") pod "efe3c20f-1ca9-46ef-92cd-de439ba78250" (UID: "efe3c20f-1ca9-46ef-92cd-de439ba78250"). InnerVolumeSpecName "kube-api-access-dmc2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.281392 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "efe3c20f-1ca9-46ef-92cd-de439ba78250" (UID: "efe3c20f-1ca9-46ef-92cd-de439ba78250"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.325593 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "efe3c20f-1ca9-46ef-92cd-de439ba78250" (UID: "efe3c20f-1ca9-46ef-92cd-de439ba78250"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.344696 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efe3c20f-1ca9-46ef-92cd-de439ba78250" (UID: "efe3c20f-1ca9-46ef-92cd-de439ba78250"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.359993 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.360038 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmc2g\" (UniqueName: \"kubernetes.io/projected/efe3c20f-1ca9-46ef-92cd-de439ba78250-kube-api-access-dmc2g\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.360051 4817 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.360063 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe3c20f-1ca9-46ef-92cd-de439ba78250-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.362845 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.460743 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-config\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461189 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dhxc\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-kube-api-access-4dhxc\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461232 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ff3653e-1322-4868-89d5-13179cd71578-prometheus-metric-storage-rulefiles-0\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461291 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ff3653e-1322-4868-89d5-13179cd71578-config-out\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461312 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-tls-assets\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461349 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-thanos-prometheus-http-client-file\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461372 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-web-config\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.461433 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"6ff3653e-1322-4868-89d5-13179cd71578\" (UID: \"6ff3653e-1322-4868-89d5-13179cd71578\") " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.464499 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-config" (OuterVolumeSpecName: "config") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.468799 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.469118 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ff3653e-1322-4868-89d5-13179cd71578-config-out" (OuterVolumeSpecName: "config-out") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.469154 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff3653e-1322-4868-89d5-13179cd71578-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.469317 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.473055 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-kube-api-access-4dhxc" (OuterVolumeSpecName: "kube-api-access-4dhxc") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "kube-api-access-4dhxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.475034 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.486194 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6df83271-9eeb-4eee-b838-feec85dde399","Type":"ContainerStarted","Data":"ec6720e81aa48c63e6dba5fd7a4b498fc9dde46cf864bd1611f85708e37e6488"} Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.492763 4817 generic.go:334] "Generic (PLEG): container finished" podID="efe3c20f-1ca9-46ef-92cd-de439ba78250" containerID="1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b" exitCode=137 Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.492837 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.492878 4817 scope.go:117] "RemoveContainer" containerID="1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.496960 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ff3653e-1322-4868-89d5-13179cd71578" containerID="f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" exitCode=0 Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.496984 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ff3653e-1322-4868-89d5-13179cd71578" containerID="822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" exitCode=0 Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.496994 4817 generic.go:334] "Generic (PLEG): container finished" podID="6ff3653e-1322-4868-89d5-13179cd71578" containerID="80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" exitCode=0 Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.497039 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerDied","Data":"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55"} Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.497061 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerDied","Data":"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c"} Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.497073 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerDied","Data":"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da"} Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.497083 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6ff3653e-1322-4868-89d5-13179cd71578","Type":"ContainerDied","Data":"68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a"} Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.497253 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.500910 4817 generic.go:334] "Generic (PLEG): container finished" podID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerID="9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76" exitCode=0 Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.501039 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerDied","Data":"9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76"} Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.502308 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-web-config" (OuterVolumeSpecName: "web-config") pod "6ff3653e-1322-4868-89d5-13179cd71578" (UID: "6ff3653e-1322-4868-89d5-13179cd71578"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.509399 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.509381531 podStartE2EDuration="3.509381531s" podCreationTimestamp="2025-11-28 15:22:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:22:37.508848637 +0000 UTC m=+3180.096827053" watchObservedRunningTime="2025-11-28 15:22:37.509381531 +0000 UTC m=+3180.097359797" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.522912 4817 scope.go:117] "RemoveContainer" containerID="1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.523631 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b\": container with ID starting with 1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b not found: ID does not exist" containerID="1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.523663 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b"} err="failed to get container status \"1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b\": rpc error: code = NotFound desc = could not find container \"1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b\": container with ID starting with 1e3b417c74d1d76f9a7f2a114ed0a1cec3595853b0ff088be5cd714756f3fa1b not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.523687 4817 scope.go:117] "RemoveContainer" containerID="f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.535297 4817 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" podUID="6df83271-9eeb-4eee-b838-feec85dde399" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.547830 4817 scope.go:117] "RemoveContainer" containerID="822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563895 4817 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6ff3653e-1322-4868-89d5-13179cd71578-config-out\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563920 4817 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563929 4817 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563938 4817 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-web-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563959 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563969 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ff3653e-1322-4868-89d5-13179cd71578-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563979 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dhxc\" (UniqueName: \"kubernetes.io/projected/6ff3653e-1322-4868-89d5-13179cd71578-kube-api-access-4dhxc\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.563988 4817 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6ff3653e-1322-4868-89d5-13179cd71578-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.567602 4817 scope.go:117] "RemoveContainer" containerID="80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.583049 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.589026 4817 scope.go:117] "RemoveContainer" containerID="a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.611514 4817 scope.go:117] "RemoveContainer" containerID="f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.611940 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": container with ID starting with f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55 not found: ID does not exist" containerID="f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.611967 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55"} err="failed to get container status \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": rpc error: code = NotFound desc = could not find container \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": container with ID starting with f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55 not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.611990 4817 scope.go:117] "RemoveContainer" containerID="822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.612347 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": container with ID starting with 822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c not found: ID does not exist" containerID="822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.612377 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c"} err="failed to get container status \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": rpc error: code = NotFound desc = could not find container \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": container with ID starting with 822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.612392 4817 scope.go:117] "RemoveContainer" containerID="80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.613032 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": container with ID starting with 80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da not found: ID does not exist" containerID="80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613068 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da"} err="failed to get container status \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": rpc error: code = NotFound desc = could not find container \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": container with ID starting with 80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613084 4817 scope.go:117] "RemoveContainer" containerID="a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.613392 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": container with ID starting with a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9 not found: ID does not exist" containerID="a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613418 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9"} err="failed to get container status \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": rpc error: code = NotFound desc = could not find container \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": container with ID starting with a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9 not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613435 4817 scope.go:117] "RemoveContainer" containerID="f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613624 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55"} err="failed to get container status \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": rpc error: code = NotFound desc = could not find container \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": container with ID starting with f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55 not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613642 4817 scope.go:117] "RemoveContainer" containerID="822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613828 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c"} err="failed to get container status \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": rpc error: code = NotFound desc = could not find container \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": container with ID starting with 822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613846 4817 scope.go:117] "RemoveContainer" containerID="80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.613998 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da"} err="failed to get container status \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": rpc error: code = NotFound desc = could not find container \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": container with ID starting with 80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.614014 4817 scope.go:117] "RemoveContainer" containerID="a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.614169 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9"} err="failed to get container status \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": rpc error: code = NotFound desc = could not find container \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": container with ID starting with a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9 not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.614187 4817 scope.go:117] "RemoveContainer" containerID="f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.614362 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55"} err="failed to get container status \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": rpc error: code = NotFound desc = could not find container \"f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55\": container with ID starting with f2b9ac572bcefb0176d05c034f7c6994785e012f37250f5cdcfa1c1a0ae71e55 not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.614379 4817 scope.go:117] "RemoveContainer" containerID="822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.615180 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c"} err="failed to get container status \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": rpc error: code = NotFound desc = could not find container \"822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c\": container with ID starting with 822128288da961e937d8c6146f8d295ab3a440ba28b38a5a42cfe9eb230d2e8c not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.615200 4817 scope.go:117] "RemoveContainer" containerID="80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.615369 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da"} err="failed to get container status \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": rpc error: code = NotFound desc = could not find container \"80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da\": container with ID starting with 80195c5bb028c9ec8f83dac9c35baad2c1dcdbe262b04fcfa7a0c191327e77da not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.615388 4817 scope.go:117] "RemoveContainer" containerID="a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.615565 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9"} err="failed to get container status \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": rpc error: code = NotFound desc = could not find container \"a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9\": container with ID starting with a3baa74dc4f2add4e7ebce733740ef1775fe697196de5c3775368a457eebc9d9 not found: ID does not exist" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.665192 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.751589 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe3c20f-1ca9-46ef-92cd-de439ba78250" path="/var/lib/kubelet/pods/efe3c20f-1ca9-46ef-92cd-de439ba78250/volumes" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.830301 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.842811 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.857674 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.858099 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="thanos-sidecar" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858117 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="thanos-sidecar" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.858144 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="prometheus" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858150 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="prometheus" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.858175 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="config-reloader" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858181 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="config-reloader" Nov 28 15:22:37 crc kubenswrapper[4817]: E1128 15:22:37.858198 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="init-config-reloader" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858204 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="init-config-reloader" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858392 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="prometheus" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858415 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="config-reloader" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.858430 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff3653e-1322-4868-89d5-13179cd71578" containerName="thanos-sidecar" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.860082 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.862562 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.863212 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.863389 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.863463 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jkmqg" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.863841 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.871126 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.878024 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.891972 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972235 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972278 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972299 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972323 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972354 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972368 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972393 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972413 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972434 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972495 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmsvx\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-kube-api-access-lmsvx\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:37 crc kubenswrapper[4817]: I1128 15:22:37.972515 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074477 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmsvx\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-kube-api-access-lmsvx\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074522 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074622 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074645 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074662 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074686 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074733 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074749 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074776 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074797 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.074818 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.076817 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.077307 4817 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.079972 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.084131 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.084593 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.086175 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.087995 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.091371 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.093558 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.094681 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.094947 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmsvx\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-kube-api-access-lmsvx\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.124606 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.189554 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.711148 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:22:38 crc kubenswrapper[4817]: W1128 15:22:38.712768 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b19e0f8_d05d_47ad_8bad_53dc2e0998c6.slice/crio-62dfb232ce4b53015a4def9110e97087a5ad92a1a4bf80aaebedd2f99ede8941 WatchSource:0}: Error finding container 62dfb232ce4b53015a4def9110e97087a5ad92a1a4bf80aaebedd2f99ede8941: Status 404 returned error can't find the container with id 62dfb232ce4b53015a4def9110e97087a5ad92a1a4bf80aaebedd2f99ede8941 Nov 28 15:22:38 crc kubenswrapper[4817]: I1128 15:22:38.737322 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:22:38 crc kubenswrapper[4817]: E1128 15:22:38.737566 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.140277 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.201467 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-internal-tls-certs\") pod \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.201559 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-scripts\") pod \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.201650 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-combined-ca-bundle\") pod \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.201667 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-config-data\") pod \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.201779 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-public-tls-certs\") pod \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.201858 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z57fl\" (UniqueName: \"kubernetes.io/projected/471e9faa-e3ae-4f2c-8894-d6a84ec97417-kube-api-access-z57fl\") pod \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\" (UID: \"471e9faa-e3ae-4f2c-8894-d6a84ec97417\") " Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.209356 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-scripts" (OuterVolumeSpecName: "scripts") pod "471e9faa-e3ae-4f2c-8894-d6a84ec97417" (UID: "471e9faa-e3ae-4f2c-8894-d6a84ec97417"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.215205 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/471e9faa-e3ae-4f2c-8894-d6a84ec97417-kube-api-access-z57fl" (OuterVolumeSpecName: "kube-api-access-z57fl") pod "471e9faa-e3ae-4f2c-8894-d6a84ec97417" (UID: "471e9faa-e3ae-4f2c-8894-d6a84ec97417"). InnerVolumeSpecName "kube-api-access-z57fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.304175 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.304213 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z57fl\" (UniqueName: \"kubernetes.io/projected/471e9faa-e3ae-4f2c-8894-d6a84ec97417-kube-api-access-z57fl\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.370253 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "471e9faa-e3ae-4f2c-8894-d6a84ec97417" (UID: "471e9faa-e3ae-4f2c-8894-d6a84ec97417"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.405873 4817 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.435619 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "471e9faa-e3ae-4f2c-8894-d6a84ec97417" (UID: "471e9faa-e3ae-4f2c-8894-d6a84ec97417"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.440833 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "471e9faa-e3ae-4f2c-8894-d6a84ec97417" (UID: "471e9faa-e3ae-4f2c-8894-d6a84ec97417"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.465235 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-config-data" (OuterVolumeSpecName: "config-data") pod "471e9faa-e3ae-4f2c-8894-d6a84ec97417" (UID: "471e9faa-e3ae-4f2c-8894-d6a84ec97417"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.507860 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.507896 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.507909 4817 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/471e9faa-e3ae-4f2c-8894-d6a84ec97417-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.548511 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerStarted","Data":"62dfb232ce4b53015a4def9110e97087a5ad92a1a4bf80aaebedd2f99ede8941"} Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.551177 4817 generic.go:334] "Generic (PLEG): container finished" podID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerID="ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5" exitCode=0 Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.551218 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerDied","Data":"ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5"} Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.551268 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"471e9faa-e3ae-4f2c-8894-d6a84ec97417","Type":"ContainerDied","Data":"7502c793d0bcbd74aea5fc85d77d60e18b6c2caf104dcac0d9a9b26509006bc2"} Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.551263 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.551283 4817 scope.go:117] "RemoveContainer" containerID="ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.569902 4817 scope.go:117] "RemoveContainer" containerID="9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.593019 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.598137 4817 scope.go:117] "RemoveContainer" containerID="a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.606697 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.622851 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.623351 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-notifier" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623377 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-notifier" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.623401 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-listener" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623409 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-listener" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.623421 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-evaluator" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623428 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-evaluator" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.623442 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-api" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623450 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-api" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623687 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-notifier" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623802 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-listener" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623827 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-evaluator" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.623854 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" containerName="aodh-api" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.626014 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.627449 4817 scope.go:117] "RemoveContainer" containerID="491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.630253 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.630294 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.630422 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.630524 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dzj58" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.632803 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.634428 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.656888 4817 scope.go:117] "RemoveContainer" containerID="ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.659026 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5\": container with ID starting with ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5 not found: ID does not exist" containerID="ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.659095 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5"} err="failed to get container status \"ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5\": rpc error: code = NotFound desc = could not find container \"ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5\": container with ID starting with ac19835217b455d85af19497e91404663b2b1efbec8956111664340d4cb71df5 not found: ID does not exist" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.659130 4817 scope.go:117] "RemoveContainer" containerID="9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.659429 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76\": container with ID starting with 9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76 not found: ID does not exist" containerID="9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.659469 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76"} err="failed to get container status \"9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76\": rpc error: code = NotFound desc = could not find container \"9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76\": container with ID starting with 9cff346de7617b09b6edad6feb4c90f455695fdb02c431d3580063b2a9f03f76 not found: ID does not exist" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.659486 4817 scope.go:117] "RemoveContainer" containerID="a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.659962 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643\": container with ID starting with a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643 not found: ID does not exist" containerID="a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.660005 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643"} err="failed to get container status \"a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643\": rpc error: code = NotFound desc = could not find container \"a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643\": container with ID starting with a9381682dde570d6f26fc2d890629dcd89bb39dc9c34e8e97bce4ab794f4b643 not found: ID does not exist" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.660023 4817 scope.go:117] "RemoveContainer" containerID="491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d" Nov 28 15:22:39 crc kubenswrapper[4817]: E1128 15:22:39.660581 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d\": container with ID starting with 491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d not found: ID does not exist" containerID="491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.660603 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d"} err="failed to get container status \"491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d\": rpc error: code = NotFound desc = could not find container \"491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d\": container with ID starting with 491fd1326ce612c898e14e924e1e59c5de8c9317b5802281687478ec5c44de8d not found: ID does not exist" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.711216 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-internal-tls-certs\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.711270 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-scripts\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.711292 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.711372 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hk9c\" (UniqueName: \"kubernetes.io/projected/0410b839-587b-4763-abca-f94c7eef60fb-kube-api-access-4hk9c\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.711422 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-config-data\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.711674 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-public-tls-certs\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.748426 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="471e9faa-e3ae-4f2c-8894-d6a84ec97417" path="/var/lib/kubelet/pods/471e9faa-e3ae-4f2c-8894-d6a84ec97417/volumes" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.749470 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff3653e-1322-4868-89d5-13179cd71578" path="/var/lib/kubelet/pods/6ff3653e-1322-4868-89d5-13179cd71578/volumes" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.813351 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-config-data\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.813486 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-public-tls-certs\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.813534 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-internal-tls-certs\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.813569 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-scripts\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.813588 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.813634 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hk9c\" (UniqueName: \"kubernetes.io/projected/0410b839-587b-4763-abca-f94c7eef60fb-kube-api-access-4hk9c\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.819350 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-public-tls-certs\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.819465 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-config-data\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.819466 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.820229 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-scripts\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.822190 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-internal-tls-certs\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.832378 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hk9c\" (UniqueName: \"kubernetes.io/projected/0410b839-587b-4763-abca-f94c7eef60fb-kube-api-access-4hk9c\") pod \"aodh-0\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " pod="openstack/aodh-0" Nov 28 15:22:39 crc kubenswrapper[4817]: I1128 15:22:39.950358 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:22:40 crc kubenswrapper[4817]: E1128 15:22:40.277939 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice/crio-68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a\": RecentStats: unable to find data in memory cache]" Nov 28 15:22:40 crc kubenswrapper[4817]: W1128 15:22:40.409876 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0410b839_587b_4763_abca_f94c7eef60fb.slice/crio-df438c46200c49f94c4761c2ebe6c0277162c9d62f6d2029e8df4e7356354a36 WatchSource:0}: Error finding container df438c46200c49f94c4761c2ebe6c0277162c9d62f6d2029e8df4e7356354a36: Status 404 returned error can't find the container with id df438c46200c49f94c4761c2ebe6c0277162c9d62f6d2029e8df4e7356354a36 Nov 28 15:22:40 crc kubenswrapper[4817]: I1128 15:22:40.418695 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:22:40 crc kubenswrapper[4817]: I1128 15:22:40.568262 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerStarted","Data":"df438c46200c49f94c4761c2ebe6c0277162c9d62f6d2029e8df4e7356354a36"} Nov 28 15:22:41 crc kubenswrapper[4817]: I1128 15:22:41.580884 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerStarted","Data":"7486bad9210f99ce7227cd8ca5c5b6a1fb03fc18713cfc143ee46bba06745780"} Nov 28 15:22:42 crc kubenswrapper[4817]: I1128 15:22:42.616073 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerStarted","Data":"334798f2fb5f83c2e33257aa10a65034973b867031ad5341754aee600558acb3"} Nov 28 15:22:42 crc kubenswrapper[4817]: I1128 15:22:42.619533 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerStarted","Data":"e2951dcbe614e1a7a8ff1f5a7c2c97cf3255a46d14f7b08f9a2dff0ee8e138a4"} Nov 28 15:22:43 crc kubenswrapper[4817]: I1128 15:22:43.637829 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerStarted","Data":"3fe680aab51639621afc0b96f298f16f4922c48e170dd6eefcd2fa9d45767dbc"} Nov 28 15:22:44 crc kubenswrapper[4817]: I1128 15:22:44.652741 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerStarted","Data":"58d1b6c19fe34a9539c37be808c334f24c6e94cd94bf58111ec69bd4cbf11dfc"} Nov 28 15:22:44 crc kubenswrapper[4817]: I1128 15:22:44.697670 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.725608961 podStartE2EDuration="5.697627925s" podCreationTimestamp="2025-11-28 15:22:39 +0000 UTC" firstStartedPulling="2025-11-28 15:22:40.412763833 +0000 UTC m=+3183.000742099" lastFinishedPulling="2025-11-28 15:22:43.384782797 +0000 UTC m=+3185.972761063" observedRunningTime="2025-11-28 15:22:44.689325045 +0000 UTC m=+3187.277303311" watchObservedRunningTime="2025-11-28 15:22:44.697627925 +0000 UTC m=+3187.285606221" Nov 28 15:22:49 crc kubenswrapper[4817]: I1128 15:22:49.711623 4817 generic.go:334] "Generic (PLEG): container finished" podID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerID="334798f2fb5f83c2e33257aa10a65034973b867031ad5341754aee600558acb3" exitCode=0 Nov 28 15:22:49 crc kubenswrapper[4817]: I1128 15:22:49.711670 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerDied","Data":"334798f2fb5f83c2e33257aa10a65034973b867031ad5341754aee600558acb3"} Nov 28 15:22:50 crc kubenswrapper[4817]: E1128 15:22:50.532553 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice/crio-68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a\": RecentStats: unable to find data in memory cache]" Nov 28 15:22:50 crc kubenswrapper[4817]: I1128 15:22:50.730307 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerStarted","Data":"2957a2ad004f5de3281873c9997cec9f3a90359af04a7bf24d955c8390a64fa7"} Nov 28 15:22:53 crc kubenswrapper[4817]: I1128 15:22:53.737180 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:22:53 crc kubenswrapper[4817]: E1128 15:22:53.737949 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:22:53 crc kubenswrapper[4817]: I1128 15:22:53.761868 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerStarted","Data":"357f2218b425fadf123af017d99f58c25c46d1bbf05e13c5f6b664d5d80c6a33"} Nov 28 15:22:53 crc kubenswrapper[4817]: I1128 15:22:53.761924 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerStarted","Data":"520168d8109a3d761e95dc7756ff9af32db90aff488b366584ece168c1cd110a"} Nov 28 15:22:53 crc kubenswrapper[4817]: I1128 15:22:53.810220 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.810199561 podStartE2EDuration="16.810199561s" podCreationTimestamp="2025-11-28 15:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:22:53.800291079 +0000 UTC m=+3196.388269345" watchObservedRunningTime="2025-11-28 15:22:53.810199561 +0000 UTC m=+3196.398177837" Nov 28 15:22:58 crc kubenswrapper[4817]: I1128 15:22:58.190152 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 28 15:23:00 crc kubenswrapper[4817]: E1128 15:23:00.773532 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice/crio-68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice\": RecentStats: unable to find data in memory cache]" Nov 28 15:23:04 crc kubenswrapper[4817]: I1128 15:23:04.738044 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:23:04 crc kubenswrapper[4817]: E1128 15:23:04.738960 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:23:08 crc kubenswrapper[4817]: I1128 15:23:08.190199 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 28 15:23:08 crc kubenswrapper[4817]: I1128 15:23:08.197436 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 28 15:23:08 crc kubenswrapper[4817]: I1128 15:23:08.938534 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 28 15:23:11 crc kubenswrapper[4817]: E1128 15:23:11.021687 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice/crio-68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a\": RecentStats: unable to find data in memory cache]" Nov 28 15:23:16 crc kubenswrapper[4817]: I1128 15:23:16.737488 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:23:16 crc kubenswrapper[4817]: E1128 15:23:16.738756 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:23:21 crc kubenswrapper[4817]: E1128 15:23:21.344181 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice/crio-68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice\": RecentStats: unable to find data in memory cache]" Nov 28 15:23:28 crc kubenswrapper[4817]: I1128 15:23:28.738255 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:23:28 crc kubenswrapper[4817]: E1128 15:23:28.739403 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:23:31 crc kubenswrapper[4817]: E1128 15:23:31.659505 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice/crio-68ade56d39d5d8d854ed4b8f53f79d08fae34617185aa6edab5a2f6b8838e67a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff3653e_1322_4868_89d5_13179cd71578.slice\": RecentStats: unable to find data in memory cache]" Nov 28 15:23:37 crc kubenswrapper[4817]: E1128 15:23:37.774256 4817 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/43efe44bb4483ac498e8c2e1b4d4b54567315b5bccce8849e7cce951bb11e1f8/diff" to get inode usage: stat /var/lib/containers/storage/overlay/43efe44bb4483ac498e8c2e1b4d4b54567315b5bccce8849e7cce951bb11e1f8/diff: no such file or directory, extraDiskErr: Nov 28 15:23:43 crc kubenswrapper[4817]: I1128 15:23:43.737204 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:23:43 crc kubenswrapper[4817]: E1128 15:23:43.738146 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:23:54 crc kubenswrapper[4817]: I1128 15:23:54.737962 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:23:54 crc kubenswrapper[4817]: E1128 15:23:54.739266 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.139252 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bb6h4"] Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.144377 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.150346 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb6h4"] Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.299266 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgldr\" (UniqueName: \"kubernetes.io/projected/07114c47-268c-4341-bde0-e9fda1c4dc88-kube-api-access-xgldr\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.299412 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-catalog-content\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.299647 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-utilities\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.402334 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-utilities\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.402572 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgldr\" (UniqueName: \"kubernetes.io/projected/07114c47-268c-4341-bde0-e9fda1c4dc88-kube-api-access-xgldr\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.402650 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-catalog-content\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.403223 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-utilities\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.403245 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-catalog-content\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.445564 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgldr\" (UniqueName: \"kubernetes.io/projected/07114c47-268c-4341-bde0-e9fda1c4dc88-kube-api-access-xgldr\") pod \"redhat-marketplace-bb6h4\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.484869 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:03 crc kubenswrapper[4817]: I1128 15:24:03.983897 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb6h4"] Nov 28 15:24:04 crc kubenswrapper[4817]: I1128 15:24:04.577642 4817 generic.go:334] "Generic (PLEG): container finished" podID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerID="398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5" exitCode=0 Nov 28 15:24:04 crc kubenswrapper[4817]: I1128 15:24:04.577746 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb6h4" event={"ID":"07114c47-268c-4341-bde0-e9fda1c4dc88","Type":"ContainerDied","Data":"398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5"} Nov 28 15:24:04 crc kubenswrapper[4817]: I1128 15:24:04.578006 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb6h4" event={"ID":"07114c47-268c-4341-bde0-e9fda1c4dc88","Type":"ContainerStarted","Data":"062c402348c80191f1fcb47de4deef07330ab9c3cb2524ac23202646fc282185"} Nov 28 15:24:05 crc kubenswrapper[4817]: I1128 15:24:05.738060 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:24:05 crc kubenswrapper[4817]: E1128 15:24:05.739005 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:24:06 crc kubenswrapper[4817]: I1128 15:24:06.610827 4817 generic.go:334] "Generic (PLEG): container finished" podID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerID="a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0" exitCode=0 Nov 28 15:24:06 crc kubenswrapper[4817]: I1128 15:24:06.610894 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb6h4" event={"ID":"07114c47-268c-4341-bde0-e9fda1c4dc88","Type":"ContainerDied","Data":"a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0"} Nov 28 15:24:07 crc kubenswrapper[4817]: I1128 15:24:07.622854 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb6h4" event={"ID":"07114c47-268c-4341-bde0-e9fda1c4dc88","Type":"ContainerStarted","Data":"f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e"} Nov 28 15:24:07 crc kubenswrapper[4817]: I1128 15:24:07.655072 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bb6h4" podStartSLOduration=2.171040571 podStartE2EDuration="4.655057235s" podCreationTimestamp="2025-11-28 15:24:03 +0000 UTC" firstStartedPulling="2025-11-28 15:24:04.580315229 +0000 UTC m=+3267.168293495" lastFinishedPulling="2025-11-28 15:24:07.064331793 +0000 UTC m=+3269.652310159" observedRunningTime="2025-11-28 15:24:07.654585353 +0000 UTC m=+3270.242563669" watchObservedRunningTime="2025-11-28 15:24:07.655057235 +0000 UTC m=+3270.243035501" Nov 28 15:24:13 crc kubenswrapper[4817]: I1128 15:24:13.485758 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:13 crc kubenswrapper[4817]: I1128 15:24:13.486218 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:13 crc kubenswrapper[4817]: I1128 15:24:13.559230 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:13 crc kubenswrapper[4817]: I1128 15:24:13.769057 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:13 crc kubenswrapper[4817]: I1128 15:24:13.846441 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb6h4"] Nov 28 15:24:14 crc kubenswrapper[4817]: I1128 15:24:14.837183 4817 scope.go:117] "RemoveContainer" containerID="8052cf05e1ff2893d30364ce1fd7349dd3c60eccd6e24f7d9ce87cc1249c3bd9" Nov 28 15:24:14 crc kubenswrapper[4817]: I1128 15:24:14.866602 4817 scope.go:117] "RemoveContainer" containerID="f00192d3b73dfcd6b0e6da9485fa02b33d0386b0c9ae8f16ac6e3384a6409b16" Nov 28 15:24:14 crc kubenswrapper[4817]: I1128 15:24:14.897168 4817 scope.go:117] "RemoveContainer" containerID="e6342f7d19900580ceafb5ceea16f4586d9af2ea0ebb4304e8d091d131d2d19d" Nov 28 15:24:14 crc kubenswrapper[4817]: I1128 15:24:14.924596 4817 scope.go:117] "RemoveContainer" containerID="2f656a9de3fcca471f43f632b6e9fb00a7e1676c2d800c633566136f16e79d8a" Nov 28 15:24:15 crc kubenswrapper[4817]: I1128 15:24:15.722345 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bb6h4" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="registry-server" containerID="cri-o://f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e" gracePeriod=2 Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.236309 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.375831 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-catalog-content\") pod \"07114c47-268c-4341-bde0-e9fda1c4dc88\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.375886 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgldr\" (UniqueName: \"kubernetes.io/projected/07114c47-268c-4341-bde0-e9fda1c4dc88-kube-api-access-xgldr\") pod \"07114c47-268c-4341-bde0-e9fda1c4dc88\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.375951 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-utilities\") pod \"07114c47-268c-4341-bde0-e9fda1c4dc88\" (UID: \"07114c47-268c-4341-bde0-e9fda1c4dc88\") " Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.376949 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-utilities" (OuterVolumeSpecName: "utilities") pod "07114c47-268c-4341-bde0-e9fda1c4dc88" (UID: "07114c47-268c-4341-bde0-e9fda1c4dc88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.384171 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07114c47-268c-4341-bde0-e9fda1c4dc88-kube-api-access-xgldr" (OuterVolumeSpecName: "kube-api-access-xgldr") pod "07114c47-268c-4341-bde0-e9fda1c4dc88" (UID: "07114c47-268c-4341-bde0-e9fda1c4dc88"). InnerVolumeSpecName "kube-api-access-xgldr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.407467 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07114c47-268c-4341-bde0-e9fda1c4dc88" (UID: "07114c47-268c-4341-bde0-e9fda1c4dc88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.477982 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.478016 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgldr\" (UniqueName: \"kubernetes.io/projected/07114c47-268c-4341-bde0-e9fda1c4dc88-kube-api-access-xgldr\") on node \"crc\" DevicePath \"\"" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.478027 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07114c47-268c-4341-bde0-e9fda1c4dc88-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.738017 4817 generic.go:334] "Generic (PLEG): container finished" podID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerID="f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e" exitCode=0 Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.738050 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb6h4" event={"ID":"07114c47-268c-4341-bde0-e9fda1c4dc88","Type":"ContainerDied","Data":"f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e"} Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.738074 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb6h4" event={"ID":"07114c47-268c-4341-bde0-e9fda1c4dc88","Type":"ContainerDied","Data":"062c402348c80191f1fcb47de4deef07330ab9c3cb2524ac23202646fc282185"} Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.738097 4817 scope.go:117] "RemoveContainer" containerID="f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.738117 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb6h4" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.774994 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb6h4"] Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.776464 4817 scope.go:117] "RemoveContainer" containerID="a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.799637 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb6h4"] Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.802074 4817 scope.go:117] "RemoveContainer" containerID="398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.848991 4817 scope.go:117] "RemoveContainer" containerID="f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e" Nov 28 15:24:16 crc kubenswrapper[4817]: E1128 15:24:16.849423 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e\": container with ID starting with f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e not found: ID does not exist" containerID="f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.849548 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e"} err="failed to get container status \"f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e\": rpc error: code = NotFound desc = could not find container \"f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e\": container with ID starting with f0ab12cf381e412c26d56c8fa3bd153e2e3cbb1ee6e15bcc9358cab3ea11b36e not found: ID does not exist" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.849681 4817 scope.go:117] "RemoveContainer" containerID="a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0" Nov 28 15:24:16 crc kubenswrapper[4817]: E1128 15:24:16.850186 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0\": container with ID starting with a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0 not found: ID does not exist" containerID="a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.850313 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0"} err="failed to get container status \"a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0\": rpc error: code = NotFound desc = could not find container \"a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0\": container with ID starting with a59933ca0fab855741dd979197e5508f096a7ade78be70f1f0fcb76017fb7ac0 not found: ID does not exist" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.850410 4817 scope.go:117] "RemoveContainer" containerID="398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5" Nov 28 15:24:16 crc kubenswrapper[4817]: E1128 15:24:16.850757 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5\": container with ID starting with 398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5 not found: ID does not exist" containerID="398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5" Nov 28 15:24:16 crc kubenswrapper[4817]: I1128 15:24:16.850783 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5"} err="failed to get container status \"398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5\": rpc error: code = NotFound desc = could not find container \"398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5\": container with ID starting with 398d768e2e4a6d350ca59d9c638505673b22bdc0313c0e2face6f8d9aa617ec5 not found: ID does not exist" Nov 28 15:24:17 crc kubenswrapper[4817]: I1128 15:24:17.751891 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" path="/var/lib/kubelet/pods/07114c47-268c-4341-bde0-e9fda1c4dc88/volumes" Nov 28 15:24:18 crc kubenswrapper[4817]: I1128 15:24:18.736536 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:24:18 crc kubenswrapper[4817]: E1128 15:24:18.736893 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:24:29 crc kubenswrapper[4817]: I1128 15:24:29.737853 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:24:29 crc kubenswrapper[4817]: E1128 15:24:29.739184 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:24:40 crc kubenswrapper[4817]: I1128 15:24:40.736356 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:24:40 crc kubenswrapper[4817]: E1128 15:24:40.737185 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:24:51 crc kubenswrapper[4817]: I1128 15:24:51.738154 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:24:51 crc kubenswrapper[4817]: E1128 15:24:51.739105 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:25:04 crc kubenswrapper[4817]: I1128 15:25:04.736928 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:25:05 crc kubenswrapper[4817]: I1128 15:25:05.267087 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"22078db55db9e8ec4497828df9cd05524a1c712fed40b1c69ef71f9624d2935a"} Nov 28 15:25:13 crc kubenswrapper[4817]: I1128 15:25:13.513574 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:25:15 crc kubenswrapper[4817]: I1128 15:25:15.987546 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:25:15 crc kubenswrapper[4817]: I1128 15:25:15.990050 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="prometheus" containerID="cri-o://2957a2ad004f5de3281873c9997cec9f3a90359af04a7bf24d955c8390a64fa7" gracePeriod=600 Nov 28 15:25:15 crc kubenswrapper[4817]: I1128 15:25:15.990128 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="thanos-sidecar" containerID="cri-o://357f2218b425fadf123af017d99f58c25c46d1bbf05e13c5f6b664d5d80c6a33" gracePeriod=600 Nov 28 15:25:15 crc kubenswrapper[4817]: I1128 15:25:15.990133 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="config-reloader" containerID="cri-o://520168d8109a3d761e95dc7756ff9af32db90aff488b366584ece168c1cd110a" gracePeriod=600 Nov 28 15:25:16 crc kubenswrapper[4817]: I1128 15:25:16.403770 4817 generic.go:334] "Generic (PLEG): container finished" podID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerID="357f2218b425fadf123af017d99f58c25c46d1bbf05e13c5f6b664d5d80c6a33" exitCode=0 Nov 28 15:25:16 crc kubenswrapper[4817]: I1128 15:25:16.404446 4817 generic.go:334] "Generic (PLEG): container finished" podID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerID="520168d8109a3d761e95dc7756ff9af32db90aff488b366584ece168c1cd110a" exitCode=0 Nov 28 15:25:16 crc kubenswrapper[4817]: I1128 15:25:16.404630 4817 generic.go:334] "Generic (PLEG): container finished" podID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerID="2957a2ad004f5de3281873c9997cec9f3a90359af04a7bf24d955c8390a64fa7" exitCode=0 Nov 28 15:25:16 crc kubenswrapper[4817]: I1128 15:25:16.403778 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerDied","Data":"357f2218b425fadf123af017d99f58c25c46d1bbf05e13c5f6b664d5d80c6a33"} Nov 28 15:25:16 crc kubenswrapper[4817]: I1128 15:25:16.404982 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerDied","Data":"520168d8109a3d761e95dc7756ff9af32db90aff488b366584ece168c1cd110a"} Nov 28 15:25:16 crc kubenswrapper[4817]: I1128 15:25:16.405185 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerDied","Data":"2957a2ad004f5de3281873c9997cec9f3a90359af04a7bf24d955c8390a64fa7"} Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.512871 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.684117 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.684198 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685196 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685383 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmsvx\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-kube-api-access-lmsvx\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685456 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-thanos-prometheus-http-client-file\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685553 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config-out\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685634 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685760 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-tls-assets\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685847 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.685937 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-prometheus-metric-storage-rulefiles-0\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.686015 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-secret-combined-ca-bundle\") pod \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\" (UID: \"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6\") " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.687745 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.691000 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config-out" (OuterVolumeSpecName: "config-out") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.692346 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config" (OuterVolumeSpecName: "config") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.692846 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.693074 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-kube-api-access-lmsvx" (OuterVolumeSpecName: "kube-api-access-lmsvx") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "kube-api-access-lmsvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.693426 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.693576 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.695102 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.701493 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.703533 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.788496 4817 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.788762 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.788844 4817 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.788925 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmsvx\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-kube-api-access-lmsvx\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.788998 4817 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.789080 4817 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-config-out\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.789179 4817 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.789261 4817 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.789367 4817 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.789445 4817 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.801808 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config" (OuterVolumeSpecName: "web-config") pod "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" (UID: "5b19e0f8-d05d-47ad-8bad-53dc2e0998c6"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.824436 4817 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.891188 4817 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6-web-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:17 crc kubenswrapper[4817]: I1128 15:25:17.891224 4817 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.439189 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5b19e0f8-d05d-47ad-8bad-53dc2e0998c6","Type":"ContainerDied","Data":"62dfb232ce4b53015a4def9110e97087a5ad92a1a4bf80aaebedd2f99ede8941"} Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.439298 4817 scope.go:117] "RemoveContainer" containerID="357f2218b425fadf123af017d99f58c25c46d1bbf05e13c5f6b664d5d80c6a33" Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.439338 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.479474 4817 scope.go:117] "RemoveContainer" containerID="520168d8109a3d761e95dc7756ff9af32db90aff488b366584ece168c1cd110a" Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.516160 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.532285 4817 scope.go:117] "RemoveContainer" containerID="2957a2ad004f5de3281873c9997cec9f3a90359af04a7bf24d955c8390a64fa7" Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.536592 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:25:18 crc kubenswrapper[4817]: I1128 15:25:18.558291 4817 scope.go:117] "RemoveContainer" containerID="334798f2fb5f83c2e33257aa10a65034973b867031ad5341754aee600558acb3" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.665571 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.666851 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="config-reloader" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.666873 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="config-reloader" Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.666896 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="prometheus" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.666908 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="prometheus" Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.666922 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="extract-utilities" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.666932 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="extract-utilities" Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.666958 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="extract-content" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.666967 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="extract-content" Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.666988 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="init-config-reloader" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.666997 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="init-config-reloader" Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.667015 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="thanos-sidecar" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.667026 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="thanos-sidecar" Nov 28 15:25:19 crc kubenswrapper[4817]: E1128 15:25:19.667132 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="registry-server" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.667146 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="registry-server" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.667452 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="prometheus" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.667481 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="thanos-sidecar" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.667500 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" containerName="config-reloader" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.667519 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="07114c47-268c-4341-bde0-e9fda1c4dc88" containerName="registry-server" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.670613 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.674471 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.675150 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.676697 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.677071 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.677091 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jkmqg" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.677294 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.683165 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.693501 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.762582 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b19e0f8-d05d-47ad-8bad-53dc2e0998c6" path="/var/lib/kubelet/pods/5b19e0f8-d05d-47ad-8bad-53dc2e0998c6/volumes" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.838361 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.838692 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.838893 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839023 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839284 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839408 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839571 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-config\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839683 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839926 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839968 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcvzl\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-kube-api-access-zcvzl\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.839993 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.941810 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.941852 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcvzl\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-kube-api-access-zcvzl\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.941877 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.941969 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942001 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942017 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942035 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942079 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942094 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942142 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-config\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942159 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.942712 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.947191 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.949547 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.949562 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.950206 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.950491 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.950996 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.951516 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.956714 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.969431 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-config\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:19 crc kubenswrapper[4817]: I1128 15:25:19.970506 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcvzl\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-kube-api-access-zcvzl\") pod \"prometheus-metric-storage-0\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:20 crc kubenswrapper[4817]: I1128 15:25:20.055081 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:20 crc kubenswrapper[4817]: I1128 15:25:20.531216 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:25:20 crc kubenswrapper[4817]: W1128 15:25:20.531872 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9706b6e_2dd0_41f6_814d_6421ef044f26.slice/crio-915b00e427a07a488e5972d4a152068b609bfe5fb1a724557ba82ca4c3c2f79d WatchSource:0}: Error finding container 915b00e427a07a488e5972d4a152068b609bfe5fb1a724557ba82ca4c3c2f79d: Status 404 returned error can't find the container with id 915b00e427a07a488e5972d4a152068b609bfe5fb1a724557ba82ca4c3c2f79d Nov 28 15:25:21 crc kubenswrapper[4817]: I1128 15:25:21.472744 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerStarted","Data":"915b00e427a07a488e5972d4a152068b609bfe5fb1a724557ba82ca4c3c2f79d"} Nov 28 15:25:25 crc kubenswrapper[4817]: I1128 15:25:25.521325 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerStarted","Data":"62bbf122eb53aa81a9ed73f6e49625c2c1c051264a7ce25132922be07274ec1e"} Nov 28 15:25:33 crc kubenswrapper[4817]: I1128 15:25:33.608656 4817 generic.go:334] "Generic (PLEG): container finished" podID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerID="62bbf122eb53aa81a9ed73f6e49625c2c1c051264a7ce25132922be07274ec1e" exitCode=0 Nov 28 15:25:33 crc kubenswrapper[4817]: I1128 15:25:33.608752 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerDied","Data":"62bbf122eb53aa81a9ed73f6e49625c2c1c051264a7ce25132922be07274ec1e"} Nov 28 15:25:34 crc kubenswrapper[4817]: I1128 15:25:34.631913 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerStarted","Data":"c26c15b1884dd84a08408e5d7cb02b4e98b492b5b2f7a6673a5f7aa374249afb"} Nov 28 15:25:38 crc kubenswrapper[4817]: I1128 15:25:38.695987 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerStarted","Data":"9637aa4362f166b3d7ebd96ff62acd1c34ffdb4789e5452d66ac17cf51fde763"} Nov 28 15:25:38 crc kubenswrapper[4817]: I1128 15:25:38.696352 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerStarted","Data":"a5acf63aa1d66b06b56b50f471f8fe5b8696b794a5642bf260ec106957fdab4d"} Nov 28 15:25:38 crc kubenswrapper[4817]: I1128 15:25:38.752227 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.752205769 podStartE2EDuration="19.752205769s" podCreationTimestamp="2025-11-28 15:25:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:25:38.738653326 +0000 UTC m=+3361.326631592" watchObservedRunningTime="2025-11-28 15:25:38.752205769 +0000 UTC m=+3361.340184035" Nov 28 15:25:40 crc kubenswrapper[4817]: I1128 15:25:40.055400 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:50 crc kubenswrapper[4817]: I1128 15:25:50.055449 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:50 crc kubenswrapper[4817]: I1128 15:25:50.065810 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 28 15:25:50 crc kubenswrapper[4817]: I1128 15:25:50.849939 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 28 15:27:00 crc kubenswrapper[4817]: I1128 15:27:00.096559 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-429b-account-create-update-vdxjt"] Nov 28 15:27:00 crc kubenswrapper[4817]: I1128 15:27:00.109054 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-429b-account-create-update-vdxjt"] Nov 28 15:27:00 crc kubenswrapper[4817]: I1128 15:27:00.120523 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-bxc9f"] Nov 28 15:27:00 crc kubenswrapper[4817]: I1128 15:27:00.130455 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-bxc9f"] Nov 28 15:27:01 crc kubenswrapper[4817]: I1128 15:27:01.773925 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09854314-9203-4849-a0c8-b3ce751451fc" path="/var/lib/kubelet/pods/09854314-9203-4849-a0c8-b3ce751451fc/volumes" Nov 28 15:27:01 crc kubenswrapper[4817]: I1128 15:27:01.775994 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c95aca32-0e83-4fb7-9715-b2c86caf6cbe" path="/var/lib/kubelet/pods/c95aca32-0e83-4fb7-9715-b2c86caf6cbe/volumes" Nov 28 15:27:12 crc kubenswrapper[4817]: I1128 15:27:12.043564 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-kx4rn"] Nov 28 15:27:12 crc kubenswrapper[4817]: I1128 15:27:12.062945 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-kx4rn"] Nov 28 15:27:13 crc kubenswrapper[4817]: I1128 15:27:13.767490 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ae908e0-75f9-407b-a4dc-6eecb4792d0c" path="/var/lib/kubelet/pods/9ae908e0-75f9-407b-a4dc-6eecb4792d0c/volumes" Nov 28 15:27:15 crc kubenswrapper[4817]: I1128 15:27:15.083324 4817 scope.go:117] "RemoveContainer" containerID="3753b95faeedff614bb23f6cb6b99edeb866342c399b2937458ad6ca8ca1f663" Nov 28 15:27:15 crc kubenswrapper[4817]: I1128 15:27:15.125356 4817 scope.go:117] "RemoveContainer" containerID="8557924642bc3e55980a94a71a4a81ddf1925065d35134e84809338cfd0684b0" Nov 28 15:27:15 crc kubenswrapper[4817]: I1128 15:27:15.178312 4817 scope.go:117] "RemoveContainer" containerID="1fd46706fa32bab4bc161bfda07359787405ab3d923a276f70572282afa6a692" Nov 28 15:27:15 crc kubenswrapper[4817]: I1128 15:27:15.481248 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.092604 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.093964 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-evaluator" containerID="cri-o://e2951dcbe614e1a7a8ff1f5a7c2c97cf3255a46d14f7b08f9a2dff0ee8e138a4" gracePeriod=30 Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.093980 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-notifier" containerID="cri-o://3fe680aab51639621afc0b96f298f16f4922c48e170dd6eefcd2fa9d45767dbc" gracePeriod=30 Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.093911 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-api" containerID="cri-o://7486bad9210f99ce7227cd8ca5c5b6a1fb03fc18713cfc143ee46bba06745780" gracePeriod=30 Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.093972 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-listener" containerID="cri-o://58d1b6c19fe34a9539c37be808c334f24c6e94cd94bf58111ec69bd4cbf11dfc" gracePeriod=30 Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.958345 4817 generic.go:334] "Generic (PLEG): container finished" podID="0410b839-587b-4763-abca-f94c7eef60fb" containerID="e2951dcbe614e1a7a8ff1f5a7c2c97cf3255a46d14f7b08f9a2dff0ee8e138a4" exitCode=0 Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.958375 4817 generic.go:334] "Generic (PLEG): container finished" podID="0410b839-587b-4763-abca-f94c7eef60fb" containerID="7486bad9210f99ce7227cd8ca5c5b6a1fb03fc18713cfc143ee46bba06745780" exitCode=0 Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.958397 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerDied","Data":"e2951dcbe614e1a7a8ff1f5a7c2c97cf3255a46d14f7b08f9a2dff0ee8e138a4"} Nov 28 15:27:17 crc kubenswrapper[4817]: I1128 15:27:17.958423 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerDied","Data":"7486bad9210f99ce7227cd8ca5c5b6a1fb03fc18713cfc143ee46bba06745780"} Nov 28 15:27:18 crc kubenswrapper[4817]: E1128 15:27:18.053922 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0410b839_587b_4763_abca_f94c7eef60fb.slice/crio-e2951dcbe614e1a7a8ff1f5a7c2c97cf3255a46d14f7b08f9a2dff0ee8e138a4.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:27:20 crc kubenswrapper[4817]: I1128 15:27:20.992977 4817 generic.go:334] "Generic (PLEG): container finished" podID="0410b839-587b-4763-abca-f94c7eef60fb" containerID="58d1b6c19fe34a9539c37be808c334f24c6e94cd94bf58111ec69bd4cbf11dfc" exitCode=0 Nov 28 15:27:20 crc kubenswrapper[4817]: I1128 15:27:20.993624 4817 generic.go:334] "Generic (PLEG): container finished" podID="0410b839-587b-4763-abca-f94c7eef60fb" containerID="3fe680aab51639621afc0b96f298f16f4922c48e170dd6eefcd2fa9d45767dbc" exitCode=0 Nov 28 15:27:20 crc kubenswrapper[4817]: I1128 15:27:20.993080 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerDied","Data":"58d1b6c19fe34a9539c37be808c334f24c6e94cd94bf58111ec69bd4cbf11dfc"} Nov 28 15:27:20 crc kubenswrapper[4817]: I1128 15:27:20.993684 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerDied","Data":"3fe680aab51639621afc0b96f298f16f4922c48e170dd6eefcd2fa9d45767dbc"} Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.259896 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.446044 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hk9c\" (UniqueName: \"kubernetes.io/projected/0410b839-587b-4763-abca-f94c7eef60fb-kube-api-access-4hk9c\") pod \"0410b839-587b-4763-abca-f94c7eef60fb\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.446331 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-scripts\") pod \"0410b839-587b-4763-abca-f94c7eef60fb\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.446377 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-internal-tls-certs\") pod \"0410b839-587b-4763-abca-f94c7eef60fb\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.446483 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-combined-ca-bundle\") pod \"0410b839-587b-4763-abca-f94c7eef60fb\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.446554 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-public-tls-certs\") pod \"0410b839-587b-4763-abca-f94c7eef60fb\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.446570 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-config-data\") pod \"0410b839-587b-4763-abca-f94c7eef60fb\" (UID: \"0410b839-587b-4763-abca-f94c7eef60fb\") " Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.451587 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0410b839-587b-4763-abca-f94c7eef60fb-kube-api-access-4hk9c" (OuterVolumeSpecName: "kube-api-access-4hk9c") pod "0410b839-587b-4763-abca-f94c7eef60fb" (UID: "0410b839-587b-4763-abca-f94c7eef60fb"). InnerVolumeSpecName "kube-api-access-4hk9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.456957 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-scripts" (OuterVolumeSpecName: "scripts") pod "0410b839-587b-4763-abca-f94c7eef60fb" (UID: "0410b839-587b-4763-abca-f94c7eef60fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.513510 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0410b839-587b-4763-abca-f94c7eef60fb" (UID: "0410b839-587b-4763-abca-f94c7eef60fb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.517641 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0410b839-587b-4763-abca-f94c7eef60fb" (UID: "0410b839-587b-4763-abca-f94c7eef60fb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.549041 4817 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.549080 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hk9c\" (UniqueName: \"kubernetes.io/projected/0410b839-587b-4763-abca-f94c7eef60fb-kube-api-access-4hk9c\") on node \"crc\" DevicePath \"\"" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.549095 4817 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.549106 4817 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.556155 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0410b839-587b-4763-abca-f94c7eef60fb" (UID: "0410b839-587b-4763-abca-f94c7eef60fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.577042 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-config-data" (OuterVolumeSpecName: "config-data") pod "0410b839-587b-4763-abca-f94c7eef60fb" (UID: "0410b839-587b-4763-abca-f94c7eef60fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.650469 4817 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 15:27:21 crc kubenswrapper[4817]: I1128 15:27:21.650507 4817 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0410b839-587b-4763-abca-f94c7eef60fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.002814 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0410b839-587b-4763-abca-f94c7eef60fb","Type":"ContainerDied","Data":"df438c46200c49f94c4761c2ebe6c0277162c9d62f6d2029e8df4e7356354a36"} Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.002865 4817 scope.go:117] "RemoveContainer" containerID="58d1b6c19fe34a9539c37be808c334f24c6e94cd94bf58111ec69bd4cbf11dfc" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.002975 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.031800 4817 scope.go:117] "RemoveContainer" containerID="3fe680aab51639621afc0b96f298f16f4922c48e170dd6eefcd2fa9d45767dbc" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.036490 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.054075 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.063893 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 28 15:27:22 crc kubenswrapper[4817]: E1128 15:27:22.064317 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-api" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064333 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-api" Nov 28 15:27:22 crc kubenswrapper[4817]: E1128 15:27:22.064347 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-listener" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064354 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-listener" Nov 28 15:27:22 crc kubenswrapper[4817]: E1128 15:27:22.064380 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-notifier" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064388 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-notifier" Nov 28 15:27:22 crc kubenswrapper[4817]: E1128 15:27:22.064400 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-evaluator" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064407 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-evaluator" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064616 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-api" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064633 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-notifier" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064650 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-evaluator" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064664 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="0410b839-587b-4763-abca-f94c7eef60fb" containerName="aodh-listener" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.064874 4817 scope.go:117] "RemoveContainer" containerID="e2951dcbe614e1a7a8ff1f5a7c2c97cf3255a46d14f7b08f9a2dff0ee8e138a4" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.066600 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.068637 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.068831 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.068890 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dzj58" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.069015 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.069113 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.090501 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.099637 4817 scope.go:117] "RemoveContainer" containerID="7486bad9210f99ce7227cd8ca5c5b6a1fb03fc18713cfc143ee46bba06745780" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.158871 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-internal-tls-certs\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.159014 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z74nc\" (UniqueName: \"kubernetes.io/projected/e80382bf-f531-4e80-9ff4-2392da846d3e-kube-api-access-z74nc\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.159086 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-scripts\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.159302 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.159347 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-public-tls-certs\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.159471 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-config-data\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.262610 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-scripts\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.262694 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.262743 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-public-tls-certs\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.262800 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-config-data\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.262833 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-internal-tls-certs\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.262913 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z74nc\" (UniqueName: \"kubernetes.io/projected/e80382bf-f531-4e80-9ff4-2392da846d3e-kube-api-access-z74nc\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.268344 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-public-tls-certs\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.268445 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-config-data\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.268889 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-internal-tls-certs\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.268939 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.270315 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e80382bf-f531-4e80-9ff4-2392da846d3e-scripts\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.280806 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z74nc\" (UniqueName: \"kubernetes.io/projected/e80382bf-f531-4e80-9ff4-2392da846d3e-kube-api-access-z74nc\") pod \"aodh-0\" (UID: \"e80382bf-f531-4e80-9ff4-2392da846d3e\") " pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.390256 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.884175 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 28 15:27:22 crc kubenswrapper[4817]: I1128 15:27:22.888708 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:27:23 crc kubenswrapper[4817]: I1128 15:27:23.012123 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e80382bf-f531-4e80-9ff4-2392da846d3e","Type":"ContainerStarted","Data":"70028bb131e205c8e48daba951516352704445f67c1211c56aee56bc33d8a92f"} Nov 28 15:27:23 crc kubenswrapper[4817]: I1128 15:27:23.750886 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0410b839-587b-4763-abca-f94c7eef60fb" path="/var/lib/kubelet/pods/0410b839-587b-4763-abca-f94c7eef60fb/volumes" Nov 28 15:27:24 crc kubenswrapper[4817]: I1128 15:27:24.025115 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e80382bf-f531-4e80-9ff4-2392da846d3e","Type":"ContainerStarted","Data":"54e7c74b8fd32f8462c279369fc562a24547be88178045aa70fb08d8b594fc01"} Nov 28 15:27:24 crc kubenswrapper[4817]: I1128 15:27:24.045267 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:27:24 crc kubenswrapper[4817]: I1128 15:27:24.045343 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:27:25 crc kubenswrapper[4817]: I1128 15:27:25.037111 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e80382bf-f531-4e80-9ff4-2392da846d3e","Type":"ContainerStarted","Data":"ab9f4b914dd6337cc8537443f69928f2352da9d82edb307bcc2aba99037f994e"} Nov 28 15:27:26 crc kubenswrapper[4817]: I1128 15:27:26.054410 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e80382bf-f531-4e80-9ff4-2392da846d3e","Type":"ContainerStarted","Data":"211f3509e66b65cd1620dc9418f4d14d186478d2d3aa8a01d6218413ee082c39"} Nov 28 15:27:27 crc kubenswrapper[4817]: I1128 15:27:27.066605 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e80382bf-f531-4e80-9ff4-2392da846d3e","Type":"ContainerStarted","Data":"33f0a84cb28c6c901608460197c2d6fea48f7c51872d3f661b538e78c1ebbdd3"} Nov 28 15:27:27 crc kubenswrapper[4817]: I1128 15:27:27.090692 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.130785023 podStartE2EDuration="5.09066882s" podCreationTimestamp="2025-11-28 15:27:22 +0000 UTC" firstStartedPulling="2025-11-28 15:27:22.888502101 +0000 UTC m=+3465.476480367" lastFinishedPulling="2025-11-28 15:27:25.848385898 +0000 UTC m=+3468.436364164" observedRunningTime="2025-11-28 15:27:27.084917135 +0000 UTC m=+3469.672895401" watchObservedRunningTime="2025-11-28 15:27:27.09066882 +0000 UTC m=+3469.678647086" Nov 28 15:27:54 crc kubenswrapper[4817]: I1128 15:27:54.045411 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:27:54 crc kubenswrapper[4817]: I1128 15:27:54.045909 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.045533 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.046123 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.046169 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.047077 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"22078db55db9e8ec4497828df9cd05524a1c712fed40b1c69ef71f9624d2935a"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.047145 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://22078db55db9e8ec4497828df9cd05524a1c712fed40b1c69ef71f9624d2935a" gracePeriod=600 Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.737263 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="22078db55db9e8ec4497828df9cd05524a1c712fed40b1c69ef71f9624d2935a" exitCode=0 Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.737352 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"22078db55db9e8ec4497828df9cd05524a1c712fed40b1c69ef71f9624d2935a"} Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.737953 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068"} Nov 28 15:28:24 crc kubenswrapper[4817]: I1128 15:28:24.737985 4817 scope.go:117] "RemoveContainer" containerID="64228625b16b44690f87680ae4d34811c0a4ea3b2a9517a3d0313dc8e13a630c" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.710002 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r98sd"] Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.715425 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.732628 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r98sd"] Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.814568 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-catalog-content\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.814666 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-utilities\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.814760 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt452\" (UniqueName: \"kubernetes.io/projected/e93e7299-4c0c-46f8-aff1-24a3e489c36f-kube-api-access-nt452\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.917584 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-utilities\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.918163 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt452\" (UniqueName: \"kubernetes.io/projected/e93e7299-4c0c-46f8-aff1-24a3e489c36f-kube-api-access-nt452\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.918308 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-catalog-content\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.918536 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-utilities\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.918813 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-catalog-content\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:38 crc kubenswrapper[4817]: I1128 15:28:38.945412 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt452\" (UniqueName: \"kubernetes.io/projected/e93e7299-4c0c-46f8-aff1-24a3e489c36f-kube-api-access-nt452\") pod \"certified-operators-r98sd\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:39 crc kubenswrapper[4817]: I1128 15:28:39.050533 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:39 crc kubenswrapper[4817]: I1128 15:28:39.584784 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r98sd"] Nov 28 15:28:39 crc kubenswrapper[4817]: I1128 15:28:39.978216 4817 generic.go:334] "Generic (PLEG): container finished" podID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerID="4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676" exitCode=0 Nov 28 15:28:39 crc kubenswrapper[4817]: I1128 15:28:39.978591 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerDied","Data":"4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676"} Nov 28 15:28:39 crc kubenswrapper[4817]: I1128 15:28:39.978664 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerStarted","Data":"9fd7b7f15fcb11caa677c81ccfb6cac463a9e7620da0ecd0f7b55df15467d058"} Nov 28 15:28:40 crc kubenswrapper[4817]: I1128 15:28:40.991001 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerStarted","Data":"4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd"} Nov 28 15:28:42 crc kubenswrapper[4817]: I1128 15:28:42.005093 4817 generic.go:334] "Generic (PLEG): container finished" podID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerID="4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd" exitCode=0 Nov 28 15:28:42 crc kubenswrapper[4817]: I1128 15:28:42.005233 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerDied","Data":"4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd"} Nov 28 15:28:43 crc kubenswrapper[4817]: I1128 15:28:43.020158 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerStarted","Data":"12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d"} Nov 28 15:28:49 crc kubenswrapper[4817]: I1128 15:28:49.051159 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:49 crc kubenswrapper[4817]: I1128 15:28:49.051872 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:49 crc kubenswrapper[4817]: I1128 15:28:49.150282 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:49 crc kubenswrapper[4817]: I1128 15:28:49.185608 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r98sd" podStartSLOduration=8.454675729 podStartE2EDuration="11.185581281s" podCreationTimestamp="2025-11-28 15:28:38 +0000 UTC" firstStartedPulling="2025-11-28 15:28:39.981768213 +0000 UTC m=+3542.569746489" lastFinishedPulling="2025-11-28 15:28:42.712673765 +0000 UTC m=+3545.300652041" observedRunningTime="2025-11-28 15:28:43.041808468 +0000 UTC m=+3545.629786754" watchObservedRunningTime="2025-11-28 15:28:49.185581281 +0000 UTC m=+3551.773559587" Nov 28 15:28:49 crc kubenswrapper[4817]: I1128 15:28:49.252049 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:49 crc kubenswrapper[4817]: I1128 15:28:49.408806 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r98sd"] Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.132530 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r98sd" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="registry-server" containerID="cri-o://12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d" gracePeriod=2 Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.594783 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.705939 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-catalog-content\") pod \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.706212 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt452\" (UniqueName: \"kubernetes.io/projected/e93e7299-4c0c-46f8-aff1-24a3e489c36f-kube-api-access-nt452\") pod \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.706257 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-utilities\") pod \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\" (UID: \"e93e7299-4c0c-46f8-aff1-24a3e489c36f\") " Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.708388 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-utilities" (OuterVolumeSpecName: "utilities") pod "e93e7299-4c0c-46f8-aff1-24a3e489c36f" (UID: "e93e7299-4c0c-46f8-aff1-24a3e489c36f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.741970 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93e7299-4c0c-46f8-aff1-24a3e489c36f-kube-api-access-nt452" (OuterVolumeSpecName: "kube-api-access-nt452") pod "e93e7299-4c0c-46f8-aff1-24a3e489c36f" (UID: "e93e7299-4c0c-46f8-aff1-24a3e489c36f"). InnerVolumeSpecName "kube-api-access-nt452". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.806847 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e93e7299-4c0c-46f8-aff1-24a3e489c36f" (UID: "e93e7299-4c0c-46f8-aff1-24a3e489c36f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.810136 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.810164 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt452\" (UniqueName: \"kubernetes.io/projected/e93e7299-4c0c-46f8-aff1-24a3e489c36f-kube-api-access-nt452\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:51 crc kubenswrapper[4817]: I1128 15:28:51.810176 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e7299-4c0c-46f8-aff1-24a3e489c36f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.149001 4817 generic.go:334] "Generic (PLEG): container finished" podID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerID="12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d" exitCode=0 Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.149081 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerDied","Data":"12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d"} Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.149135 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r98sd" event={"ID":"e93e7299-4c0c-46f8-aff1-24a3e489c36f","Type":"ContainerDied","Data":"9fd7b7f15fcb11caa677c81ccfb6cac463a9e7620da0ecd0f7b55df15467d058"} Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.149153 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r98sd" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.149168 4817 scope.go:117] "RemoveContainer" containerID="12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.185569 4817 scope.go:117] "RemoveContainer" containerID="4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.215619 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r98sd"] Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.227220 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r98sd"] Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.232303 4817 scope.go:117] "RemoveContainer" containerID="4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.290425 4817 scope.go:117] "RemoveContainer" containerID="12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d" Nov 28 15:28:52 crc kubenswrapper[4817]: E1128 15:28:52.290917 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d\": container with ID starting with 12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d not found: ID does not exist" containerID="12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.290952 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d"} err="failed to get container status \"12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d\": rpc error: code = NotFound desc = could not find container \"12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d\": container with ID starting with 12003f8548d3f4cbb6d4744c372dbe56265762e4c9ecc224f7df9d4ba7e56f6d not found: ID does not exist" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.290975 4817 scope.go:117] "RemoveContainer" containerID="4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd" Nov 28 15:28:52 crc kubenswrapper[4817]: E1128 15:28:52.291246 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd\": container with ID starting with 4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd not found: ID does not exist" containerID="4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.291270 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd"} err="failed to get container status \"4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd\": rpc error: code = NotFound desc = could not find container \"4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd\": container with ID starting with 4ef633d3695d1389d088acde55f5334f579413e51c61c470d703e02a592832bd not found: ID does not exist" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.291284 4817 scope.go:117] "RemoveContainer" containerID="4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676" Nov 28 15:28:52 crc kubenswrapper[4817]: E1128 15:28:52.291592 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676\": container with ID starting with 4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676 not found: ID does not exist" containerID="4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676" Nov 28 15:28:52 crc kubenswrapper[4817]: I1128 15:28:52.291616 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676"} err="failed to get container status \"4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676\": rpc error: code = NotFound desc = could not find container \"4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676\": container with ID starting with 4b6147a5ce76136848db344fc48c24f1b1dc273080d706cdbabd1f13d306b676 not found: ID does not exist" Nov 28 15:28:53 crc kubenswrapper[4817]: I1128 15:28:53.754565 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" path="/var/lib/kubelet/pods/e93e7299-4c0c-46f8-aff1-24a3e489c36f/volumes" Nov 28 15:29:17 crc kubenswrapper[4817]: I1128 15:29:17.489657 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:29:21 crc kubenswrapper[4817]: I1128 15:29:21.676600 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:29:21 crc kubenswrapper[4817]: I1128 15:29:21.677449 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="prometheus" containerID="cri-o://c26c15b1884dd84a08408e5d7cb02b4e98b492b5b2f7a6673a5f7aa374249afb" gracePeriod=600 Nov 28 15:29:21 crc kubenswrapper[4817]: I1128 15:29:21.677901 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="thanos-sidecar" containerID="cri-o://9637aa4362f166b3d7ebd96ff62acd1c34ffdb4789e5452d66ac17cf51fde763" gracePeriod=600 Nov 28 15:29:21 crc kubenswrapper[4817]: I1128 15:29:21.677944 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="config-reloader" containerID="cri-o://a5acf63aa1d66b06b56b50f471f8fe5b8696b794a5642bf260ec106957fdab4d" gracePeriod=600 Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.538369 4817 generic.go:334] "Generic (PLEG): container finished" podID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerID="9637aa4362f166b3d7ebd96ff62acd1c34ffdb4789e5452d66ac17cf51fde763" exitCode=0 Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.538590 4817 generic.go:334] "Generic (PLEG): container finished" podID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerID="a5acf63aa1d66b06b56b50f471f8fe5b8696b794a5642bf260ec106957fdab4d" exitCode=0 Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.538598 4817 generic.go:334] "Generic (PLEG): container finished" podID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerID="c26c15b1884dd84a08408e5d7cb02b4e98b492b5b2f7a6673a5f7aa374249afb" exitCode=0 Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.538452 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerDied","Data":"9637aa4362f166b3d7ebd96ff62acd1c34ffdb4789e5452d66ac17cf51fde763"} Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.538640 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerDied","Data":"a5acf63aa1d66b06b56b50f471f8fe5b8696b794a5642bf260ec106957fdab4d"} Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.538655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerDied","Data":"c26c15b1884dd84a08408e5d7cb02b4e98b492b5b2f7a6673a5f7aa374249afb"} Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.680942 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.852307 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-tls-assets\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.852365 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-thanos-prometheus-http-client-file\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.852408 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-db\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.852465 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-config-out\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.852483 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853266 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-db" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "prometheus-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853358 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-rulefiles-0\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853406 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcvzl\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-kube-api-access-zcvzl\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853444 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-config\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853472 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853506 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.853550 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-secret-combined-ca-bundle\") pod \"d9706b6e-2dd0-41f6-814d-6421ef044f26\" (UID: \"d9706b6e-2dd0-41f6-814d-6421ef044f26\") " Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.854635 4817 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-db\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.854973 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.858938 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-config" (OuterVolumeSpecName: "config") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.860935 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.861770 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.862776 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-kube-api-access-zcvzl" (OuterVolumeSpecName: "kube-api-access-zcvzl") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "kube-api-access-zcvzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.864968 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.865002 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.865371 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.870941 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-config-out" (OuterVolumeSpecName: "config-out") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.938430 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config" (OuterVolumeSpecName: "web-config") pod "d9706b6e-2dd0-41f6-814d-6421ef044f26" (UID: "d9706b6e-2dd0-41f6-814d-6421ef044f26"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.956544 4817 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.956710 4817 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.956816 4817 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d9706b6e-2dd0-41f6-814d-6421ef044f26-config-out\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.956968 4817 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.957054 4817 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d9706b6e-2dd0-41f6-814d-6421ef044f26-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.957133 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcvzl\" (UniqueName: \"kubernetes.io/projected/d9706b6e-2dd0-41f6-814d-6421ef044f26-kube-api-access-zcvzl\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.957218 4817 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.957364 4817 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.957466 4817 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-web-config\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:22 crc kubenswrapper[4817]: I1128 15:29:22.957562 4817 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9706b6e-2dd0-41f6-814d-6421ef044f26-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.553298 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d9706b6e-2dd0-41f6-814d-6421ef044f26","Type":"ContainerDied","Data":"915b00e427a07a488e5972d4a152068b609bfe5fb1a724557ba82ca4c3c2f79d"} Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.553356 4817 scope.go:117] "RemoveContainer" containerID="9637aa4362f166b3d7ebd96ff62acd1c34ffdb4789e5452d66ac17cf51fde763" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.553510 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.581194 4817 scope.go:117] "RemoveContainer" containerID="a5acf63aa1d66b06b56b50f471f8fe5b8696b794a5642bf260ec106957fdab4d" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.608185 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.620490 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.626483 4817 scope.go:117] "RemoveContainer" containerID="c26c15b1884dd84a08408e5d7cb02b4e98b492b5b2f7a6673a5f7aa374249afb" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645145 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645531 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="registry-server" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645547 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="registry-server" Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645572 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="prometheus" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645578 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="prometheus" Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645591 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="extract-content" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645598 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="extract-content" Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645610 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="init-config-reloader" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645616 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="init-config-reloader" Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645628 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="extract-utilities" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645634 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="extract-utilities" Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645646 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="thanos-sidecar" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645651 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="thanos-sidecar" Nov 28 15:29:23 crc kubenswrapper[4817]: E1128 15:29:23.645662 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="config-reloader" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645668 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="config-reloader" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645854 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="config-reloader" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645871 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93e7299-4c0c-46f8-aff1-24a3e489c36f" containerName="registry-server" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645883 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="prometheus" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.645896 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" containerName="thanos-sidecar" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.647520 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.651419 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.651710 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jkmqg" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.651991 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.652845 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.653318 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.653872 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.662651 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.674258 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.683895 4817 scope.go:117] "RemoveContainer" containerID="62bbf122eb53aa81a9ed73f6e49625c2c1c051264a7ce25132922be07274ec1e" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.754932 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9706b6e-2dd0-41f6-814d-6421ef044f26" path="/var/lib/kubelet/pods/d9706b6e-2dd0-41f6-814d-6421ef044f26/volumes" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.795913 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.795985 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796015 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796150 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796238 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796505 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xpnq\" (UniqueName: \"kubernetes.io/projected/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-kube-api-access-9xpnq\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796561 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-config\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796590 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796618 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796637 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.796742 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904410 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xpnq\" (UniqueName: \"kubernetes.io/projected/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-kube-api-access-9xpnq\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904488 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-config\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904521 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904554 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904580 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904667 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904817 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904896 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904937 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.904961 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.905011 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.917613 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.919300 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.927147 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.927596 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.933606 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.943322 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.944450 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.945006 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-config\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.945991 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.949167 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.972521 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xpnq\" (UniqueName: \"kubernetes.io/projected/7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f-kube-api-access-9xpnq\") pod \"prometheus-metric-storage-0\" (UID: \"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f\") " pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:23 crc kubenswrapper[4817]: I1128 15:29:23.982666 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:24 crc kubenswrapper[4817]: I1128 15:29:24.578658 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 28 15:29:25 crc kubenswrapper[4817]: I1128 15:29:25.584624 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f","Type":"ContainerStarted","Data":"3ef4ec0e2fc09a578c8b733701115c8e6658c96aff2b0e9a72d4a63187800aff"} Nov 28 15:29:29 crc kubenswrapper[4817]: I1128 15:29:29.651099 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f","Type":"ContainerStarted","Data":"1818fb335ac7ed67eaa10e67d815b11c1e74cd8ca5da4efaffa6f431105fb2c4"} Nov 28 15:29:39 crc kubenswrapper[4817]: I1128 15:29:39.802961 4817 generic.go:334] "Generic (PLEG): container finished" podID="7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f" containerID="1818fb335ac7ed67eaa10e67d815b11c1e74cd8ca5da4efaffa6f431105fb2c4" exitCode=0 Nov 28 15:29:39 crc kubenswrapper[4817]: I1128 15:29:39.803090 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f","Type":"ContainerDied","Data":"1818fb335ac7ed67eaa10e67d815b11c1e74cd8ca5da4efaffa6f431105fb2c4"} Nov 28 15:29:40 crc kubenswrapper[4817]: I1128 15:29:40.822496 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f","Type":"ContainerStarted","Data":"a5034518fdc2537e39ff679f13cd8d7ba7fc79254c272356509ad3017ed74c38"} Nov 28 15:29:44 crc kubenswrapper[4817]: I1128 15:29:44.872898 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f","Type":"ContainerStarted","Data":"d987ab154bec671070dcaefb809f102ca80efb45d9887417b569b925b0b9f449"} Nov 28 15:29:44 crc kubenswrapper[4817]: I1128 15:29:44.873489 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f","Type":"ContainerStarted","Data":"65cda96ee37b806a9b13581ccd93999162a377f39e7029aa1a3afa5e81a44bdd"} Nov 28 15:29:44 crc kubenswrapper[4817]: I1128 15:29:44.914294 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.914276823 podStartE2EDuration="21.914276823s" podCreationTimestamp="2025-11-28 15:29:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:29:44.91254259 +0000 UTC m=+3607.500520886" watchObservedRunningTime="2025-11-28 15:29:44.914276823 +0000 UTC m=+3607.502255089" Nov 28 15:29:48 crc kubenswrapper[4817]: I1128 15:29:48.983617 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:53 crc kubenswrapper[4817]: I1128 15:29:53.983764 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:53 crc kubenswrapper[4817]: I1128 15:29:53.992072 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 28 15:29:54 crc kubenswrapper[4817]: I1128 15:29:54.985630 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.190032 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g"] Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.192657 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.195426 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.195819 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.206040 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g"] Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.336945 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j9rj\" (UniqueName: \"kubernetes.io/projected/5681446b-21e7-4022-852c-7e665d477e3c-kube-api-access-2j9rj\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.336991 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5681446b-21e7-4022-852c-7e665d477e3c-config-volume\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.337036 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5681446b-21e7-4022-852c-7e665d477e3c-secret-volume\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.439082 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j9rj\" (UniqueName: \"kubernetes.io/projected/5681446b-21e7-4022-852c-7e665d477e3c-kube-api-access-2j9rj\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.439130 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5681446b-21e7-4022-852c-7e665d477e3c-config-volume\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.439175 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5681446b-21e7-4022-852c-7e665d477e3c-secret-volume\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.440349 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5681446b-21e7-4022-852c-7e665d477e3c-config-volume\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.447626 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5681446b-21e7-4022-852c-7e665d477e3c-secret-volume\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.462424 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j9rj\" (UniqueName: \"kubernetes.io/projected/5681446b-21e7-4022-852c-7e665d477e3c-kube-api-access-2j9rj\") pod \"collect-profiles-29405730-vtd9g\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.528590 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:00 crc kubenswrapper[4817]: I1128 15:30:00.999067 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g"] Nov 28 15:30:01 crc kubenswrapper[4817]: W1128 15:30:01.003621 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5681446b_21e7_4022_852c_7e665d477e3c.slice/crio-9c0f8b0f8af09eeb5f07142c00003d190ae53829ed0342f0fe44245c2e30376d WatchSource:0}: Error finding container 9c0f8b0f8af09eeb5f07142c00003d190ae53829ed0342f0fe44245c2e30376d: Status 404 returned error can't find the container with id 9c0f8b0f8af09eeb5f07142c00003d190ae53829ed0342f0fe44245c2e30376d Nov 28 15:30:01 crc kubenswrapper[4817]: I1128 15:30:01.063573 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" event={"ID":"5681446b-21e7-4022-852c-7e665d477e3c","Type":"ContainerStarted","Data":"9c0f8b0f8af09eeb5f07142c00003d190ae53829ed0342f0fe44245c2e30376d"} Nov 28 15:30:02 crc kubenswrapper[4817]: I1128 15:30:02.078924 4817 generic.go:334] "Generic (PLEG): container finished" podID="5681446b-21e7-4022-852c-7e665d477e3c" containerID="06f32905bfb3eeab605fa16cb76444ea6033a119230e8652bf04c32dcd0501dc" exitCode=0 Nov 28 15:30:02 crc kubenswrapper[4817]: I1128 15:30:02.079020 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" event={"ID":"5681446b-21e7-4022-852c-7e665d477e3c","Type":"ContainerDied","Data":"06f32905bfb3eeab605fa16cb76444ea6033a119230e8652bf04c32dcd0501dc"} Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.470084 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.609283 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5681446b-21e7-4022-852c-7e665d477e3c-config-volume\") pod \"5681446b-21e7-4022-852c-7e665d477e3c\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.609453 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j9rj\" (UniqueName: \"kubernetes.io/projected/5681446b-21e7-4022-852c-7e665d477e3c-kube-api-access-2j9rj\") pod \"5681446b-21e7-4022-852c-7e665d477e3c\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.609762 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5681446b-21e7-4022-852c-7e665d477e3c-secret-volume\") pod \"5681446b-21e7-4022-852c-7e665d477e3c\" (UID: \"5681446b-21e7-4022-852c-7e665d477e3c\") " Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.610258 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5681446b-21e7-4022-852c-7e665d477e3c-config-volume" (OuterVolumeSpecName: "config-volume") pod "5681446b-21e7-4022-852c-7e665d477e3c" (UID: "5681446b-21e7-4022-852c-7e665d477e3c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.610469 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5681446b-21e7-4022-852c-7e665d477e3c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.618843 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5681446b-21e7-4022-852c-7e665d477e3c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5681446b-21e7-4022-852c-7e665d477e3c" (UID: "5681446b-21e7-4022-852c-7e665d477e3c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.619652 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5681446b-21e7-4022-852c-7e665d477e3c-kube-api-access-2j9rj" (OuterVolumeSpecName: "kube-api-access-2j9rj") pod "5681446b-21e7-4022-852c-7e665d477e3c" (UID: "5681446b-21e7-4022-852c-7e665d477e3c"). InnerVolumeSpecName "kube-api-access-2j9rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.713349 4817 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5681446b-21e7-4022-852c-7e665d477e3c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:03 crc kubenswrapper[4817]: I1128 15:30:03.713854 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j9rj\" (UniqueName: \"kubernetes.io/projected/5681446b-21e7-4022-852c-7e665d477e3c-kube-api-access-2j9rj\") on node \"crc\" DevicePath \"\"" Nov 28 15:30:04 crc kubenswrapper[4817]: I1128 15:30:04.116206 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" event={"ID":"5681446b-21e7-4022-852c-7e665d477e3c","Type":"ContainerDied","Data":"9c0f8b0f8af09eeb5f07142c00003d190ae53829ed0342f0fe44245c2e30376d"} Nov 28 15:30:04 crc kubenswrapper[4817]: I1128 15:30:04.116264 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c0f8b0f8af09eeb5f07142c00003d190ae53829ed0342f0fe44245c2e30376d" Nov 28 15:30:04 crc kubenswrapper[4817]: I1128 15:30:04.116306 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405730-vtd9g" Nov 28 15:30:04 crc kubenswrapper[4817]: I1128 15:30:04.560341 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd"] Nov 28 15:30:04 crc kubenswrapper[4817]: I1128 15:30:04.572390 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405685-drswd"] Nov 28 15:30:05 crc kubenswrapper[4817]: I1128 15:30:05.761089 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e59a9f-fb61-4364-ace0-8940f22583cb" path="/var/lib/kubelet/pods/75e59a9f-fb61-4364-ace0-8940f22583cb/volumes" Nov 28 15:30:15 crc kubenswrapper[4817]: I1128 15:30:15.447130 4817 scope.go:117] "RemoveContainer" containerID="301e989706465d265ad00f51473a5f18b7d012a3e8f35245c2c2ac58a32f27db" Nov 28 15:30:20 crc kubenswrapper[4817]: I1128 15:30:20.374875 4817 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-s4gdr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 15:30:20 crc kubenswrapper[4817]: I1128 15:30:20.375702 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" podUID="8b50b274-12b2-4ed3-ad0a-81365aeff72c" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 15:30:20 crc kubenswrapper[4817]: I1128 15:30:20.374945 4817 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-s4gdr container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 15:30:20 crc kubenswrapper[4817]: I1128 15:30:20.376270 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s4gdr" podUID="8b50b274-12b2-4ed3-ad0a-81365aeff72c" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 15:30:24 crc kubenswrapper[4817]: I1128 15:30:24.045586 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:30:24 crc kubenswrapper[4817]: I1128 15:30:24.046185 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.694833 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ptmmq"] Nov 28 15:30:52 crc kubenswrapper[4817]: E1128 15:30:52.695996 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5681446b-21e7-4022-852c-7e665d477e3c" containerName="collect-profiles" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.696015 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="5681446b-21e7-4022-852c-7e665d477e3c" containerName="collect-profiles" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.696277 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="5681446b-21e7-4022-852c-7e665d477e3c" containerName="collect-profiles" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.697855 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.703437 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ptmmq"] Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.814744 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hktcq\" (UniqueName: \"kubernetes.io/projected/44852b24-3631-470a-b030-3534dea74f64-kube-api-access-hktcq\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.815213 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-catalog-content\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.815341 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-utilities\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.917107 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hktcq\" (UniqueName: \"kubernetes.io/projected/44852b24-3631-470a-b030-3534dea74f64-kube-api-access-hktcq\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.917235 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-catalog-content\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.917301 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-utilities\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.917917 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-utilities\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.918014 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-catalog-content\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:52 crc kubenswrapper[4817]: I1128 15:30:52.956426 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hktcq\" (UniqueName: \"kubernetes.io/projected/44852b24-3631-470a-b030-3534dea74f64-kube-api-access-hktcq\") pod \"redhat-operators-ptmmq\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:53 crc kubenswrapper[4817]: I1128 15:30:53.022418 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:30:53 crc kubenswrapper[4817]: I1128 15:30:53.338348 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ptmmq"] Nov 28 15:30:53 crc kubenswrapper[4817]: I1128 15:30:53.974644 4817 generic.go:334] "Generic (PLEG): container finished" podID="44852b24-3631-470a-b030-3534dea74f64" containerID="c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9" exitCode=0 Nov 28 15:30:53 crc kubenswrapper[4817]: I1128 15:30:53.974818 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerDied","Data":"c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9"} Nov 28 15:30:53 crc kubenswrapper[4817]: I1128 15:30:53.975049 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerStarted","Data":"843ab7749b7b47f2dce4dac5ced6c9fc4d330898ac2d8f1912855efb64de3e41"} Nov 28 15:30:54 crc kubenswrapper[4817]: I1128 15:30:54.044708 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:30:54 crc kubenswrapper[4817]: I1128 15:30:54.044835 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:30:56 crc kubenswrapper[4817]: I1128 15:30:56.000271 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerStarted","Data":"afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22"} Nov 28 15:30:57 crc kubenswrapper[4817]: I1128 15:30:57.014082 4817 generic.go:334] "Generic (PLEG): container finished" podID="44852b24-3631-470a-b030-3534dea74f64" containerID="afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22" exitCode=0 Nov 28 15:30:57 crc kubenswrapper[4817]: I1128 15:30:57.014158 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerDied","Data":"afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22"} Nov 28 15:30:59 crc kubenswrapper[4817]: I1128 15:30:59.040588 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerStarted","Data":"85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d"} Nov 28 15:30:59 crc kubenswrapper[4817]: I1128 15:30:59.065635 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ptmmq" podStartSLOduration=3.277622638 podStartE2EDuration="7.065617372s" podCreationTimestamp="2025-11-28 15:30:52 +0000 UTC" firstStartedPulling="2025-11-28 15:30:53.989825151 +0000 UTC m=+3676.577803417" lastFinishedPulling="2025-11-28 15:30:57.777819885 +0000 UTC m=+3680.365798151" observedRunningTime="2025-11-28 15:30:59.059282173 +0000 UTC m=+3681.647260479" watchObservedRunningTime="2025-11-28 15:30:59.065617372 +0000 UTC m=+3681.653595628" Nov 28 15:31:03 crc kubenswrapper[4817]: I1128 15:31:03.023112 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:31:03 crc kubenswrapper[4817]: I1128 15:31:03.023858 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:31:04 crc kubenswrapper[4817]: I1128 15:31:04.095205 4817 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ptmmq" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="registry-server" probeResult="failure" output=< Nov 28 15:31:04 crc kubenswrapper[4817]: timeout: failed to connect service ":50051" within 1s Nov 28 15:31:04 crc kubenswrapper[4817]: > Nov 28 15:31:13 crc kubenswrapper[4817]: I1128 15:31:13.107599 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:31:13 crc kubenswrapper[4817]: I1128 15:31:13.174256 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:31:13 crc kubenswrapper[4817]: I1128 15:31:13.362157 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ptmmq"] Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.230269 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ptmmq" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="registry-server" containerID="cri-o://85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d" gracePeriod=2 Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.709614 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.798738 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-catalog-content\") pod \"44852b24-3631-470a-b030-3534dea74f64\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.798929 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hktcq\" (UniqueName: \"kubernetes.io/projected/44852b24-3631-470a-b030-3534dea74f64-kube-api-access-hktcq\") pod \"44852b24-3631-470a-b030-3534dea74f64\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.799074 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-utilities\") pod \"44852b24-3631-470a-b030-3534dea74f64\" (UID: \"44852b24-3631-470a-b030-3534dea74f64\") " Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.800592 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-utilities" (OuterVolumeSpecName: "utilities") pod "44852b24-3631-470a-b030-3534dea74f64" (UID: "44852b24-3631-470a-b030-3534dea74f64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.805096 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44852b24-3631-470a-b030-3534dea74f64-kube-api-access-hktcq" (OuterVolumeSpecName: "kube-api-access-hktcq") pod "44852b24-3631-470a-b030-3534dea74f64" (UID: "44852b24-3631-470a-b030-3534dea74f64"). InnerVolumeSpecName "kube-api-access-hktcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.901967 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hktcq\" (UniqueName: \"kubernetes.io/projected/44852b24-3631-470a-b030-3534dea74f64-kube-api-access-hktcq\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.902315 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:14 crc kubenswrapper[4817]: I1128 15:31:14.919435 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44852b24-3631-470a-b030-3534dea74f64" (UID: "44852b24-3631-470a-b030-3534dea74f64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.004360 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44852b24-3631-470a-b030-3534dea74f64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.241880 4817 generic.go:334] "Generic (PLEG): container finished" podID="44852b24-3631-470a-b030-3534dea74f64" containerID="85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d" exitCode=0 Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.241935 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerDied","Data":"85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d"} Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.241968 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptmmq" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.241991 4817 scope.go:117] "RemoveContainer" containerID="85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.241977 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptmmq" event={"ID":"44852b24-3631-470a-b030-3534dea74f64","Type":"ContainerDied","Data":"843ab7749b7b47f2dce4dac5ced6c9fc4d330898ac2d8f1912855efb64de3e41"} Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.273920 4817 scope.go:117] "RemoveContainer" containerID="afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.284567 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ptmmq"] Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.294689 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ptmmq"] Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.302004 4817 scope.go:117] "RemoveContainer" containerID="c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.351173 4817 scope.go:117] "RemoveContainer" containerID="85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d" Nov 28 15:31:15 crc kubenswrapper[4817]: E1128 15:31:15.351940 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d\": container with ID starting with 85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d not found: ID does not exist" containerID="85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.352012 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d"} err="failed to get container status \"85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d\": rpc error: code = NotFound desc = could not find container \"85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d\": container with ID starting with 85f802c90d3ba985be601debd9b008b502142573230e72069e52a828deefda8d not found: ID does not exist" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.352053 4817 scope.go:117] "RemoveContainer" containerID="afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22" Nov 28 15:31:15 crc kubenswrapper[4817]: E1128 15:31:15.352580 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22\": container with ID starting with afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22 not found: ID does not exist" containerID="afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.352621 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22"} err="failed to get container status \"afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22\": rpc error: code = NotFound desc = could not find container \"afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22\": container with ID starting with afab8772043fffc7166e55ecd7cc67f68417b985ba988a79b4d7172b2b5d1e22 not found: ID does not exist" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.352652 4817 scope.go:117] "RemoveContainer" containerID="c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9" Nov 28 15:31:15 crc kubenswrapper[4817]: E1128 15:31:15.352995 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9\": container with ID starting with c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9 not found: ID does not exist" containerID="c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.353026 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9"} err="failed to get container status \"c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9\": rpc error: code = NotFound desc = could not find container \"c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9\": container with ID starting with c2ddc66d5502b0fbf3211585ae2adaecda5c3b0380715bac2b4872a3d95786a9 not found: ID does not exist" Nov 28 15:31:15 crc kubenswrapper[4817]: I1128 15:31:15.766495 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44852b24-3631-470a-b030-3534dea74f64" path="/var/lib/kubelet/pods/44852b24-3631-470a-b030-3534dea74f64/volumes" Nov 28 15:31:21 crc kubenswrapper[4817]: I1128 15:31:21.247531 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.046111 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.046510 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.046567 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.047625 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.047782 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" gracePeriod=600 Nov 28 15:31:24 crc kubenswrapper[4817]: E1128 15:31:24.194294 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.357854 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" exitCode=0 Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.357927 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068"} Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.357979 4817 scope.go:117] "RemoveContainer" containerID="22078db55db9e8ec4497828df9cd05524a1c712fed40b1c69ef71f9624d2935a" Nov 28 15:31:24 crc kubenswrapper[4817]: I1128 15:31:24.360254 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:31:24 crc kubenswrapper[4817]: E1128 15:31:24.363572 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:31:37 crc kubenswrapper[4817]: I1128 15:31:37.745177 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:31:37 crc kubenswrapper[4817]: E1128 15:31:37.746635 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.305637 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fxmv8/must-gather-tdmd6"] Nov 28 15:31:40 crc kubenswrapper[4817]: E1128 15:31:40.306579 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="registry-server" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.306593 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="registry-server" Nov 28 15:31:40 crc kubenswrapper[4817]: E1128 15:31:40.306623 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="extract-content" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.306631 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="extract-content" Nov 28 15:31:40 crc kubenswrapper[4817]: E1128 15:31:40.306646 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="extract-utilities" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.306656 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="extract-utilities" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.306882 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="44852b24-3631-470a-b030-3534dea74f64" containerName="registry-server" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.308097 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.312907 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-fxmv8"/"default-dockercfg-7c46g" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.312947 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-fxmv8"/"kube-root-ca.crt" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.312906 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-fxmv8"/"openshift-service-ca.crt" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.333705 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-fxmv8/must-gather-tdmd6"] Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.461262 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr5gw\" (UniqueName: \"kubernetes.io/projected/ebf5c67c-81e6-4878-a23c-9f45e14f6697-kube-api-access-wr5gw\") pod \"must-gather-tdmd6\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.461301 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ebf5c67c-81e6-4878-a23c-9f45e14f6697-must-gather-output\") pod \"must-gather-tdmd6\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.562748 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr5gw\" (UniqueName: \"kubernetes.io/projected/ebf5c67c-81e6-4878-a23c-9f45e14f6697-kube-api-access-wr5gw\") pod \"must-gather-tdmd6\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.562811 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ebf5c67c-81e6-4878-a23c-9f45e14f6697-must-gather-output\") pod \"must-gather-tdmd6\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.563413 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ebf5c67c-81e6-4878-a23c-9f45e14f6697-must-gather-output\") pod \"must-gather-tdmd6\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.586887 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr5gw\" (UniqueName: \"kubernetes.io/projected/ebf5c67c-81e6-4878-a23c-9f45e14f6697-kube-api-access-wr5gw\") pod \"must-gather-tdmd6\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:40 crc kubenswrapper[4817]: I1128 15:31:40.627885 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:31:41 crc kubenswrapper[4817]: I1128 15:31:41.162981 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-fxmv8/must-gather-tdmd6"] Nov 28 15:31:41 crc kubenswrapper[4817]: W1128 15:31:41.172980 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf5c67c_81e6_4878_a23c_9f45e14f6697.slice/crio-5c48ef155edec4f63723355b17e2d2618b48801b389569a735cde86c15e17ff8 WatchSource:0}: Error finding container 5c48ef155edec4f63723355b17e2d2618b48801b389569a735cde86c15e17ff8: Status 404 returned error can't find the container with id 5c48ef155edec4f63723355b17e2d2618b48801b389569a735cde86c15e17ff8 Nov 28 15:31:41 crc kubenswrapper[4817]: I1128 15:31:41.556114 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" event={"ID":"ebf5c67c-81e6-4878-a23c-9f45e14f6697","Type":"ContainerStarted","Data":"5c48ef155edec4f63723355b17e2d2618b48801b389569a735cde86c15e17ff8"} Nov 28 15:31:46 crc kubenswrapper[4817]: I1128 15:31:46.608309 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" event={"ID":"ebf5c67c-81e6-4878-a23c-9f45e14f6697","Type":"ContainerStarted","Data":"e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68"} Nov 28 15:31:47 crc kubenswrapper[4817]: I1128 15:31:47.620076 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" event={"ID":"ebf5c67c-81e6-4878-a23c-9f45e14f6697","Type":"ContainerStarted","Data":"94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6"} Nov 28 15:31:47 crc kubenswrapper[4817]: I1128 15:31:47.658789 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" podStartSLOduration=2.652557696 podStartE2EDuration="7.658762957s" podCreationTimestamp="2025-11-28 15:31:40 +0000 UTC" firstStartedPulling="2025-11-28 15:31:41.175413982 +0000 UTC m=+3723.763392268" lastFinishedPulling="2025-11-28 15:31:46.181619253 +0000 UTC m=+3728.769597529" observedRunningTime="2025-11-28 15:31:47.635223686 +0000 UTC m=+3730.223201942" watchObservedRunningTime="2025-11-28 15:31:47.658762957 +0000 UTC m=+3730.246741223" Nov 28 15:31:50 crc kubenswrapper[4817]: E1128 15:31:50.400352 4817 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.236:39062->38.102.83.236:36927: write tcp 38.102.83.236:39062->38.102.83.236:36927: write: broken pipe Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.822462 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fxmv8/crc-debug-fpfz4"] Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.823692 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.889057 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngds4\" (UniqueName: \"kubernetes.io/projected/42a03c28-b0e0-41e8-8476-3c566f748027-kube-api-access-ngds4\") pod \"crc-debug-fpfz4\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.889236 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42a03c28-b0e0-41e8-8476-3c566f748027-host\") pod \"crc-debug-fpfz4\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.990323 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42a03c28-b0e0-41e8-8476-3c566f748027-host\") pod \"crc-debug-fpfz4\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.990411 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngds4\" (UniqueName: \"kubernetes.io/projected/42a03c28-b0e0-41e8-8476-3c566f748027-kube-api-access-ngds4\") pod \"crc-debug-fpfz4\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:50 crc kubenswrapper[4817]: I1128 15:31:50.990468 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42a03c28-b0e0-41e8-8476-3c566f748027-host\") pod \"crc-debug-fpfz4\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:51 crc kubenswrapper[4817]: I1128 15:31:51.015237 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngds4\" (UniqueName: \"kubernetes.io/projected/42a03c28-b0e0-41e8-8476-3c566f748027-kube-api-access-ngds4\") pod \"crc-debug-fpfz4\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:51 crc kubenswrapper[4817]: I1128 15:31:51.141552 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:31:51 crc kubenswrapper[4817]: W1128 15:31:51.174434 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42a03c28_b0e0_41e8_8476_3c566f748027.slice/crio-c4b31ac3b1c4e9b212219e73e52f73eb09d90dd297be6412e197ca471af1f4b2 WatchSource:0}: Error finding container c4b31ac3b1c4e9b212219e73e52f73eb09d90dd297be6412e197ca471af1f4b2: Status 404 returned error can't find the container with id c4b31ac3b1c4e9b212219e73e52f73eb09d90dd297be6412e197ca471af1f4b2 Nov 28 15:31:51 crc kubenswrapper[4817]: I1128 15:31:51.658540 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" event={"ID":"42a03c28-b0e0-41e8-8476-3c566f748027","Type":"ContainerStarted","Data":"c4b31ac3b1c4e9b212219e73e52f73eb09d90dd297be6412e197ca471af1f4b2"} Nov 28 15:31:51 crc kubenswrapper[4817]: I1128 15:31:51.746138 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:31:51 crc kubenswrapper[4817]: E1128 15:31:51.750206 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:32:02 crc kubenswrapper[4817]: I1128 15:32:02.738487 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:32:02 crc kubenswrapper[4817]: E1128 15:32:02.739793 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:32:07 crc kubenswrapper[4817]: I1128 15:32:07.840485 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" event={"ID":"42a03c28-b0e0-41e8-8476-3c566f748027","Type":"ContainerStarted","Data":"25f50de32384def93bf3c1878ae2ad10e365233d731c5c3ac129c58c0d27d097"} Nov 28 15:32:07 crc kubenswrapper[4817]: I1128 15:32:07.855468 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" podStartSLOduration=2.130809137 podStartE2EDuration="17.855451911s" podCreationTimestamp="2025-11-28 15:31:50 +0000 UTC" firstStartedPulling="2025-11-28 15:31:51.176818989 +0000 UTC m=+3733.764797255" lastFinishedPulling="2025-11-28 15:32:06.901461763 +0000 UTC m=+3749.489440029" observedRunningTime="2025-11-28 15:32:07.852691992 +0000 UTC m=+3750.440670268" watchObservedRunningTime="2025-11-28 15:32:07.855451911 +0000 UTC m=+3750.443430177" Nov 28 15:32:13 crc kubenswrapper[4817]: I1128 15:32:13.736617 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:32:13 crc kubenswrapper[4817]: E1128 15:32:13.737402 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.585733 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qgc92"] Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.588675 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.608281 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgc92"] Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.705917 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-utilities\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.706116 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5rpx\" (UniqueName: \"kubernetes.io/projected/db7f71ec-3b13-40be-9c77-bf1783abfb80-kube-api-access-x5rpx\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.706139 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-catalog-content\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.807948 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5rpx\" (UniqueName: \"kubernetes.io/projected/db7f71ec-3b13-40be-9c77-bf1783abfb80-kube-api-access-x5rpx\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.808019 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-catalog-content\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.808163 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-utilities\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.809012 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-utilities\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.809814 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-catalog-content\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.852397 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5rpx\" (UniqueName: \"kubernetes.io/projected/db7f71ec-3b13-40be-9c77-bf1783abfb80-kube-api-access-x5rpx\") pod \"community-operators-qgc92\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:16 crc kubenswrapper[4817]: I1128 15:32:16.929207 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:17 crc kubenswrapper[4817]: I1128 15:32:17.534372 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgc92"] Nov 28 15:32:17 crc kubenswrapper[4817]: W1128 15:32:17.565865 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb7f71ec_3b13_40be_9c77_bf1783abfb80.slice/crio-45341e31230688c7e9324bf16898ec1d23e3f82f14147951ac7b68cb62c32a3b WatchSource:0}: Error finding container 45341e31230688c7e9324bf16898ec1d23e3f82f14147951ac7b68cb62c32a3b: Status 404 returned error can't find the container with id 45341e31230688c7e9324bf16898ec1d23e3f82f14147951ac7b68cb62c32a3b Nov 28 15:32:17 crc kubenswrapper[4817]: I1128 15:32:17.938915 4817 generic.go:334] "Generic (PLEG): container finished" podID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerID="9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4" exitCode=0 Nov 28 15:32:17 crc kubenswrapper[4817]: I1128 15:32:17.939254 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerDied","Data":"9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4"} Nov 28 15:32:17 crc kubenswrapper[4817]: I1128 15:32:17.939326 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerStarted","Data":"45341e31230688c7e9324bf16898ec1d23e3f82f14147951ac7b68cb62c32a3b"} Nov 28 15:32:20 crc kubenswrapper[4817]: I1128 15:32:20.969555 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerStarted","Data":"310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5"} Nov 28 15:32:24 crc kubenswrapper[4817]: I1128 15:32:24.006196 4817 generic.go:334] "Generic (PLEG): container finished" podID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerID="310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5" exitCode=0 Nov 28 15:32:24 crc kubenswrapper[4817]: I1128 15:32:24.006246 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerDied","Data":"310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5"} Nov 28 15:32:24 crc kubenswrapper[4817]: I1128 15:32:24.010800 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:32:25 crc kubenswrapper[4817]: I1128 15:32:25.016835 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerStarted","Data":"30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3"} Nov 28 15:32:25 crc kubenswrapper[4817]: I1128 15:32:25.020343 4817 generic.go:334] "Generic (PLEG): container finished" podID="42a03c28-b0e0-41e8-8476-3c566f748027" containerID="25f50de32384def93bf3c1878ae2ad10e365233d731c5c3ac129c58c0d27d097" exitCode=0 Nov 28 15:32:25 crc kubenswrapper[4817]: I1128 15:32:25.020376 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" event={"ID":"42a03c28-b0e0-41e8-8476-3c566f748027","Type":"ContainerDied","Data":"25f50de32384def93bf3c1878ae2ad10e365233d731c5c3ac129c58c0d27d097"} Nov 28 15:32:25 crc kubenswrapper[4817]: I1128 15:32:25.039065 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qgc92" podStartSLOduration=2.481979818 podStartE2EDuration="9.039047737s" podCreationTimestamp="2025-11-28 15:32:16 +0000 UTC" firstStartedPulling="2025-11-28 15:32:17.941807164 +0000 UTC m=+3760.529785430" lastFinishedPulling="2025-11-28 15:32:24.498875083 +0000 UTC m=+3767.086853349" observedRunningTime="2025-11-28 15:32:25.036884733 +0000 UTC m=+3767.624863009" watchObservedRunningTime="2025-11-28 15:32:25.039047737 +0000 UTC m=+3767.627026003" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.130421 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.168536 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fxmv8/crc-debug-fpfz4"] Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.180569 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fxmv8/crc-debug-fpfz4"] Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.316502 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngds4\" (UniqueName: \"kubernetes.io/projected/42a03c28-b0e0-41e8-8476-3c566f748027-kube-api-access-ngds4\") pod \"42a03c28-b0e0-41e8-8476-3c566f748027\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.316553 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42a03c28-b0e0-41e8-8476-3c566f748027-host\") pod \"42a03c28-b0e0-41e8-8476-3c566f748027\" (UID: \"42a03c28-b0e0-41e8-8476-3c566f748027\") " Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.316727 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/42a03c28-b0e0-41e8-8476-3c566f748027-host" (OuterVolumeSpecName: "host") pod "42a03c28-b0e0-41e8-8476-3c566f748027" (UID: "42a03c28-b0e0-41e8-8476-3c566f748027"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.317231 4817 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42a03c28-b0e0-41e8-8476-3c566f748027-host\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.325035 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a03c28-b0e0-41e8-8476-3c566f748027-kube-api-access-ngds4" (OuterVolumeSpecName: "kube-api-access-ngds4") pod "42a03c28-b0e0-41e8-8476-3c566f748027" (UID: "42a03c28-b0e0-41e8-8476-3c566f748027"). InnerVolumeSpecName "kube-api-access-ngds4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.418864 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngds4\" (UniqueName: \"kubernetes.io/projected/42a03c28-b0e0-41e8-8476-3c566f748027-kube-api-access-ngds4\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.929454 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:26 crc kubenswrapper[4817]: I1128 15:32:26.929495 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.038985 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4b31ac3b1c4e9b212219e73e52f73eb09d90dd297be6412e197ca471af1f4b2" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.039058 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-fpfz4" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.442453 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fxmv8/crc-debug-6z6xw"] Nov 28 15:32:27 crc kubenswrapper[4817]: E1128 15:32:27.443188 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a03c28-b0e0-41e8-8476-3c566f748027" containerName="container-00" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.443214 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a03c28-b0e0-41e8-8476-3c566f748027" containerName="container-00" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.443576 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a03c28-b0e0-41e8-8476-3c566f748027" containerName="container-00" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.444704 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.506537 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.544296 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvlgc\" (UniqueName: \"kubernetes.io/projected/7b91a795-5e03-4b0f-9eba-2a57a0c38078-kube-api-access-xvlgc\") pod \"crc-debug-6z6xw\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.544768 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b91a795-5e03-4b0f-9eba-2a57a0c38078-host\") pod \"crc-debug-6z6xw\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.646261 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b91a795-5e03-4b0f-9eba-2a57a0c38078-host\") pod \"crc-debug-6z6xw\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.646403 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvlgc\" (UniqueName: \"kubernetes.io/projected/7b91a795-5e03-4b0f-9eba-2a57a0c38078-kube-api-access-xvlgc\") pod \"crc-debug-6z6xw\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.646687 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b91a795-5e03-4b0f-9eba-2a57a0c38078-host\") pod \"crc-debug-6z6xw\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.671931 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvlgc\" (UniqueName: \"kubernetes.io/projected/7b91a795-5e03-4b0f-9eba-2a57a0c38078-kube-api-access-xvlgc\") pod \"crc-debug-6z6xw\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.743593 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:32:27 crc kubenswrapper[4817]: E1128 15:32:27.744116 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.747382 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a03c28-b0e0-41e8-8476-3c566f748027" path="/var/lib/kubelet/pods/42a03c28-b0e0-41e8-8476-3c566f748027/volumes" Nov 28 15:32:27 crc kubenswrapper[4817]: I1128 15:32:27.780762 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:28 crc kubenswrapper[4817]: I1128 15:32:28.048335 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" event={"ID":"7b91a795-5e03-4b0f-9eba-2a57a0c38078","Type":"ContainerStarted","Data":"47ebef6e89c8f7c6b967b4d50ffb5b28a82436f42bb52db91e33ee6b0cb38a52"} Nov 28 15:32:29 crc kubenswrapper[4817]: I1128 15:32:29.059140 4817 generic.go:334] "Generic (PLEG): container finished" podID="7b91a795-5e03-4b0f-9eba-2a57a0c38078" containerID="0f2674e7da37830b425c16f89fb50f4557eb52809f133a17253847b2e5518dde" exitCode=1 Nov 28 15:32:29 crc kubenswrapper[4817]: I1128 15:32:29.059201 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" event={"ID":"7b91a795-5e03-4b0f-9eba-2a57a0c38078","Type":"ContainerDied","Data":"0f2674e7da37830b425c16f89fb50f4557eb52809f133a17253847b2e5518dde"} Nov 28 15:32:29 crc kubenswrapper[4817]: I1128 15:32:29.100621 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fxmv8/crc-debug-6z6xw"] Nov 28 15:32:29 crc kubenswrapper[4817]: I1128 15:32:29.109927 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fxmv8/crc-debug-6z6xw"] Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.200269 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.399862 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b91a795-5e03-4b0f-9eba-2a57a0c38078-host\") pod \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.399992 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b91a795-5e03-4b0f-9eba-2a57a0c38078-host" (OuterVolumeSpecName: "host") pod "7b91a795-5e03-4b0f-9eba-2a57a0c38078" (UID: "7b91a795-5e03-4b0f-9eba-2a57a0c38078"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.400089 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvlgc\" (UniqueName: \"kubernetes.io/projected/7b91a795-5e03-4b0f-9eba-2a57a0c38078-kube-api-access-xvlgc\") pod \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\" (UID: \"7b91a795-5e03-4b0f-9eba-2a57a0c38078\") " Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.400877 4817 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b91a795-5e03-4b0f-9eba-2a57a0c38078-host\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.409129 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b91a795-5e03-4b0f-9eba-2a57a0c38078-kube-api-access-xvlgc" (OuterVolumeSpecName: "kube-api-access-xvlgc") pod "7b91a795-5e03-4b0f-9eba-2a57a0c38078" (UID: "7b91a795-5e03-4b0f-9eba-2a57a0c38078"). InnerVolumeSpecName "kube-api-access-xvlgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:30 crc kubenswrapper[4817]: I1128 15:32:30.502514 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvlgc\" (UniqueName: \"kubernetes.io/projected/7b91a795-5e03-4b0f-9eba-2a57a0c38078-kube-api-access-xvlgc\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:31 crc kubenswrapper[4817]: I1128 15:32:31.086023 4817 scope.go:117] "RemoveContainer" containerID="0f2674e7da37830b425c16f89fb50f4557eb52809f133a17253847b2e5518dde" Nov 28 15:32:31 crc kubenswrapper[4817]: I1128 15:32:31.086165 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/crc-debug-6z6xw" Nov 28 15:32:31 crc kubenswrapper[4817]: I1128 15:32:31.761521 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b91a795-5e03-4b0f-9eba-2a57a0c38078" path="/var/lib/kubelet/pods/7b91a795-5e03-4b0f-9eba-2a57a0c38078/volumes" Nov 28 15:32:36 crc kubenswrapper[4817]: I1128 15:32:36.979599 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.042666 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgc92"] Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.171457 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qgc92" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="registry-server" containerID="cri-o://30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3" gracePeriod=2 Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.702567 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.847541 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-catalog-content\") pod \"db7f71ec-3b13-40be-9c77-bf1783abfb80\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.847785 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-utilities\") pod \"db7f71ec-3b13-40be-9c77-bf1783abfb80\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.847904 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5rpx\" (UniqueName: \"kubernetes.io/projected/db7f71ec-3b13-40be-9c77-bf1783abfb80-kube-api-access-x5rpx\") pod \"db7f71ec-3b13-40be-9c77-bf1783abfb80\" (UID: \"db7f71ec-3b13-40be-9c77-bf1783abfb80\") " Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.848780 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-utilities" (OuterVolumeSpecName: "utilities") pod "db7f71ec-3b13-40be-9c77-bf1783abfb80" (UID: "db7f71ec-3b13-40be-9c77-bf1783abfb80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.853191 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7f71ec-3b13-40be-9c77-bf1783abfb80-kube-api-access-x5rpx" (OuterVolumeSpecName: "kube-api-access-x5rpx") pod "db7f71ec-3b13-40be-9c77-bf1783abfb80" (UID: "db7f71ec-3b13-40be-9c77-bf1783abfb80"). InnerVolumeSpecName "kube-api-access-x5rpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.914644 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db7f71ec-3b13-40be-9c77-bf1783abfb80" (UID: "db7f71ec-3b13-40be-9c77-bf1783abfb80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.950461 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.950504 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5rpx\" (UniqueName: \"kubernetes.io/projected/db7f71ec-3b13-40be-9c77-bf1783abfb80-kube-api-access-x5rpx\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:37 crc kubenswrapper[4817]: I1128 15:32:37.950517 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7f71ec-3b13-40be-9c77-bf1783abfb80-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.184901 4817 generic.go:334] "Generic (PLEG): container finished" podID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerID="30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3" exitCode=0 Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.184968 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerDied","Data":"30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3"} Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.185022 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgc92" event={"ID":"db7f71ec-3b13-40be-9c77-bf1783abfb80","Type":"ContainerDied","Data":"45341e31230688c7e9324bf16898ec1d23e3f82f14147951ac7b68cb62c32a3b"} Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.185044 4817 scope.go:117] "RemoveContainer" containerID="30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.185287 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgc92" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.218545 4817 scope.go:117] "RemoveContainer" containerID="310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.223662 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgc92"] Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.232511 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qgc92"] Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.242944 4817 scope.go:117] "RemoveContainer" containerID="9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.283386 4817 scope.go:117] "RemoveContainer" containerID="30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3" Nov 28 15:32:38 crc kubenswrapper[4817]: E1128 15:32:38.283819 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3\": container with ID starting with 30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3 not found: ID does not exist" containerID="30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.283864 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3"} err="failed to get container status \"30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3\": rpc error: code = NotFound desc = could not find container \"30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3\": container with ID starting with 30bcc9ba1d7cdb1b99be4a26a109281af43318ae24cd75c56d96b5ba599012c3 not found: ID does not exist" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.283895 4817 scope.go:117] "RemoveContainer" containerID="310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5" Nov 28 15:32:38 crc kubenswrapper[4817]: E1128 15:32:38.284279 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5\": container with ID starting with 310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5 not found: ID does not exist" containerID="310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.284307 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5"} err="failed to get container status \"310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5\": rpc error: code = NotFound desc = could not find container \"310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5\": container with ID starting with 310312d44d1f421b2fd9d2d3c3d8058bd78cb1e6ef522c77e5fe4e0da7de02a5 not found: ID does not exist" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.284325 4817 scope.go:117] "RemoveContainer" containerID="9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4" Nov 28 15:32:38 crc kubenswrapper[4817]: E1128 15:32:38.284593 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4\": container with ID starting with 9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4 not found: ID does not exist" containerID="9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4" Nov 28 15:32:38 crc kubenswrapper[4817]: I1128 15:32:38.284635 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4"} err="failed to get container status \"9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4\": rpc error: code = NotFound desc = could not find container \"9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4\": container with ID starting with 9062149648ea0223565972a3a24f513867c85f0500142f0972d80565ef5ca1e4 not found: ID does not exist" Nov 28 15:32:39 crc kubenswrapper[4817]: I1128 15:32:39.752685 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" path="/var/lib/kubelet/pods/db7f71ec-3b13-40be-9c77-bf1783abfb80/volumes" Nov 28 15:32:40 crc kubenswrapper[4817]: I1128 15:32:40.737209 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:32:40 crc kubenswrapper[4817]: E1128 15:32:40.737811 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:32:54 crc kubenswrapper[4817]: I1128 15:32:54.738832 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:32:54 crc kubenswrapper[4817]: E1128 15:32:54.739560 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:33:08 crc kubenswrapper[4817]: I1128 15:33:08.738045 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:33:08 crc kubenswrapper[4817]: E1128 15:33:08.739004 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.267083 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/init-config-reloader/0.log" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.433151 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/init-config-reloader/0.log" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.446681 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/alertmanager/0.log" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.469247 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/config-reloader/0.log" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.907256 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-evaluator/0.log" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.923882 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-api/0.log" Nov 28 15:33:18 crc kubenswrapper[4817]: I1128 15:33:18.978854 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-listener/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.116681 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fb9577b84-z8dsl_15698a16-9f4e-4552-b729-27969466063d/barbican-api/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.123242 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-notifier/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.203506 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fb9577b84-z8dsl_15698a16-9f4e-4552-b729-27969466063d/barbican-api-log/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.312328 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7dd97866bb-qxt2v_7cc501c6-50dd-4f39-b19b-20f10e6b9997/barbican-keystone-listener/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.412061 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7dd97866bb-qxt2v_7cc501c6-50dd-4f39-b19b-20f10e6b9997/barbican-keystone-listener-log/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.565660 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84d84dc879-8hskz_9022c280-17b2-48fc-b818-73797815a819/barbican-worker/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.570748 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84d84dc879-8hskz_9022c280-17b2-48fc-b818-73797815a819/barbican-worker-log/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.785741 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q_3200157e-779d-485c-803e-1b0d092c02bb/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.802321 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/ceilometer-central-agent/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.873996 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/ceilometer-notification-agent/0.log" Nov 28 15:33:19 crc kubenswrapper[4817]: I1128 15:33:19.984979 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/sg-core/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.007169 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/proxy-httpd/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.092938 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8b2cd5c-abbf-4cca-98d0-e0f3db266421/cinder-api/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.149833 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8b2cd5c-abbf-4cca-98d0-e0f3db266421/cinder-api-log/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.312164 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7570a53b-a405-45ee-ab15-ab5f575bb90b/cinder-scheduler/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.446020 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7570a53b-a405-45ee-ab15-ab5f575bb90b/probe/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.691754 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8_1a488bc2-89d2-49a6-88a3-0656eb5deb56/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.736976 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:33:20 crc kubenswrapper[4817]: E1128 15:33:20.737255 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.759310 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5_18ae828a-395c-4300-9460-a063a0322335/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:20 crc kubenswrapper[4817]: I1128 15:33:20.862808 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-f8k7g_6ecb4997-6b98-4f7b-ad43-e02295ffe723/init/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.062353 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-f8k7g_6ecb4997-6b98-4f7b-ad43-e02295ffe723/init/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.170636 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q_0452060d-61d9-49c6-b16c-522ded9cb4e8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.170981 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-f8k7g_6ecb4997-6b98-4f7b-ad43-e02295ffe723/dnsmasq-dns/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.322821 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c4f615c7-7cc5-4339-8bdc-e2974582e564/glance-httpd/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.397101 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c4f615c7-7cc5-4339-8bdc-e2974582e564/glance-log/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.467391 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d0451aa8-c32f-46e7-a9d8-574cb7005894/glance-httpd/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.500865 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d0451aa8-c32f-46e7-a9d8-574cb7005894/glance-log/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.896619 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7585599b46-rxrsf_bde8e8ec-30e7-4913-83bb-2babaaf11f9a/heat-engine/0.log" Nov 28 15:33:21 crc kubenswrapper[4817]: I1128 15:33:21.922997 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6f9fcf49b7-fdlpq_2b12bb76-1c8c-4ea3-afd6-b1ff967ae174/heat-api/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.011598 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5696cfb998-r25ks_11d0830f-fa4e-4204-b32d-3ef9553325f1/heat-cfnapi/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.015819 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd_c817bef2-9ff5-4f69-8a73-1f36128879a6/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.122534 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zd8vk_62cd0765-6d4e-4e36-96df-8e3cc4bf8a91/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.337027 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c874c964f-4j6hm_b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d/keystone-api/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.408865 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29405701-7q48x_9ef2895d-d3df-4a6b-abad-c7d1bceb0871/keystone-cron/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.416710 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_95d4474e-82dd-405a-a280-726ce4f894dc/kube-state-metrics/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.640627 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm_58e1259c-59ec-4bee-867b-b90c25667b33/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.866315 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cc96c44df-rqchh_0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9/neutron-httpd/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.916166 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cc96c44df-rqchh_0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9/neutron-api/0.log" Nov 28 15:33:22 crc kubenswrapper[4817]: I1128 15:33:22.980495 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c_f7494198-71fe-4460-a0a5-0737e7fe57dd/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:23 crc kubenswrapper[4817]: I1128 15:33:23.286041 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eb11cfb0-8f52-4361-bbce-b70613a51ca6/nova-api-log/0.log" Nov 28 15:33:23 crc kubenswrapper[4817]: I1128 15:33:23.449131 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_98dff079-658b-4cb4-aa29-42eb2c020705/nova-cell0-conductor-conductor/0.log" Nov 28 15:33:23 crc kubenswrapper[4817]: I1128 15:33:23.562271 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eb11cfb0-8f52-4361-bbce-b70613a51ca6/nova-api-api/0.log" Nov 28 15:33:23 crc kubenswrapper[4817]: I1128 15:33:23.596273 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_37355c0f-d765-472c-b5d6-beb57bcf9869/nova-cell1-conductor-conductor/0.log" Nov 28 15:33:23 crc kubenswrapper[4817]: I1128 15:33:23.783959 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_514053b0-d172-46fb-ac7c-26bd118d83f0/nova-cell1-novncproxy-novncproxy/0.log" Nov 28 15:33:23 crc kubenswrapper[4817]: I1128 15:33:23.911588 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-g4mz8_8833fd86-9ef4-441d-ada9-bb24bec33153/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:24 crc kubenswrapper[4817]: I1128 15:33:24.293532 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_99283929-00ab-45d4-92c9-73cbb67f205f/nova-metadata-log/0.log" Nov 28 15:33:24 crc kubenswrapper[4817]: I1128 15:33:24.554784 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1fcffa08-235a-4101-bb4f-8933dc9102e3/nova-scheduler-scheduler/0.log" Nov 28 15:33:24 crc kubenswrapper[4817]: I1128 15:33:24.558173 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d/mysql-bootstrap/0.log" Nov 28 15:33:24 crc kubenswrapper[4817]: I1128 15:33:24.735109 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d/mysql-bootstrap/0.log" Nov 28 15:33:24 crc kubenswrapper[4817]: I1128 15:33:24.782397 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d/galera/0.log" Nov 28 15:33:24 crc kubenswrapper[4817]: I1128 15:33:24.914361 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bd723f67-725a-4942-87f2-c646eddcd172/mysql-bootstrap/0.log" Nov 28 15:33:25 crc kubenswrapper[4817]: I1128 15:33:25.391403 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_99283929-00ab-45d4-92c9-73cbb67f205f/nova-metadata-metadata/0.log" Nov 28 15:33:25 crc kubenswrapper[4817]: I1128 15:33:25.604768 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bd723f67-725a-4942-87f2-c646eddcd172/mysql-bootstrap/0.log" Nov 28 15:33:25 crc kubenswrapper[4817]: I1128 15:33:25.676772 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bd723f67-725a-4942-87f2-c646eddcd172/galera/0.log" Nov 28 15:33:25 crc kubenswrapper[4817]: I1128 15:33:25.714907 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6df83271-9eeb-4eee-b838-feec85dde399/openstackclient/0.log" Nov 28 15:33:25 crc kubenswrapper[4817]: I1128 15:33:25.909308 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9frjn_9ac085c7-55fa-4dd9-9ab6-bfd720716c60/openstack-network-exporter/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.003236 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovsdb-server-init/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.188908 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovsdb-server/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.191590 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovs-vswitchd/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.237537 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovsdb-server-init/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.479863 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pwgz2_0f58e7f4-f831-4f54-91b6-a5e97bf5152c/ovn-controller/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.484977 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-d6ls6_c49884f3-a2a8-49d5-8d6c-c2d7485909ed/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.663454 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_693b7740-a0c9-4430-941a-3f849c254803/ovn-northd/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.710880 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_693b7740-a0c9-4430-941a-3f849c254803/openstack-network-exporter/0.log" Nov 28 15:33:26 crc kubenswrapper[4817]: I1128 15:33:26.775363 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e3b52366-bcd8-4685-9f36-2adaea4e6cc4/openstack-network-exporter/0.log" Nov 28 15:33:27 crc kubenswrapper[4817]: I1128 15:33:27.443581 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ff9385f1-3c29-48be-bc80-806380c08646/openstack-network-exporter/0.log" Nov 28 15:33:27 crc kubenswrapper[4817]: I1128 15:33:27.445918 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ff9385f1-3c29-48be-bc80-806380c08646/ovsdbserver-sb/0.log" Nov 28 15:33:27 crc kubenswrapper[4817]: I1128 15:33:27.459574 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e3b52366-bcd8-4685-9f36-2adaea4e6cc4/ovsdbserver-nb/0.log" Nov 28 15:33:27 crc kubenswrapper[4817]: I1128 15:33:27.739878 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6746568d76-zwnfd_4164fdcc-c963-4a08-9bd5-2c569d096180/placement-api/0.log" Nov 28 15:33:27 crc kubenswrapper[4817]: I1128 15:33:27.834503 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6746568d76-zwnfd_4164fdcc-c963-4a08-9bd5-2c569d096180/placement-log/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.001325 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/init-config-reloader/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.090113 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/init-config-reloader/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.090284 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/config-reloader/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.112964 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/prometheus/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.208767 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/thanos-sidecar/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.296573 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a5cbe25b-ed99-4853-a3bd-726f25343f6a/setup-container/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.732360 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a5cbe25b-ed99-4853-a3bd-726f25343f6a/setup-container/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.742396 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a5cbe25b-ed99-4853-a3bd-726f25343f6a/rabbitmq/0.log" Nov 28 15:33:28 crc kubenswrapper[4817]: I1128 15:33:28.831263 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8a6a65e3-df90-4500-8deb-1dc661954e49/setup-container/0.log" Nov 28 15:33:29 crc kubenswrapper[4817]: I1128 15:33:29.049574 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8a6a65e3-df90-4500-8deb-1dc661954e49/setup-container/0.log" Nov 28 15:33:29 crc kubenswrapper[4817]: I1128 15:33:29.088736 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m_c36cc3fe-b1aa-4858-8c72-1fa223352e09/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:29 crc kubenswrapper[4817]: I1128 15:33:29.297244 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zzxtk_14f43d11-6db6-461c-9279-22fdd3e286b8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:29 crc kubenswrapper[4817]: I1128 15:33:29.495687 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd_02fe616c-2dd3-412b-a8e1-9af3150fc92c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:29 crc kubenswrapper[4817]: I1128 15:33:29.629588 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pvgld_0b6cd940-c79b-4bd7-80ff-69bb850709ae/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:29 crc kubenswrapper[4817]: I1128 15:33:29.796337 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-jzgbr_3867da3b-518a-4aa7-80b4-5d6e5f4e7dad/ssh-known-hosts-edpm-deployment/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.050312 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85df6b4fc9-7c5nt_72c26e96-3946-44a8-bb20-7786cba96b07/proxy-server/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.116162 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85df6b4fc9-7c5nt_72c26e96-3946-44a8-bb20-7786cba96b07/proxy-httpd/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.239518 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-fzhb6_f61aa2a2-1ba8-4319-ad02-2c05d6a9f662/swift-ring-rebalance/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.368543 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-auditor/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.466351 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-reaper/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.570189 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-server/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.587580 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-replicator/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.676119 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-auditor/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.753746 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8a6a65e3-df90-4500-8deb-1dc661954e49/rabbitmq/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.801533 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-replicator/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.820888 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-server/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.877309 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-updater/0.log" Nov 28 15:33:30 crc kubenswrapper[4817]: I1128 15:33:30.938451 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-auditor/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.008274 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-expirer/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.032504 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-replicator/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.113928 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-server/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.197971 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-updater/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.261428 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/rsync/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.303022 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/swift-recon-cron/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.465071 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-622cj_f9a2d436-9de0-46c8-8a53-073060648992/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:31 crc kubenswrapper[4817]: I1128 15:33:31.574708 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8_aa1fe5c5-0d23-47fb-990d-480cd18970b8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:33:33 crc kubenswrapper[4817]: I1128 15:33:33.777049 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:33:33 crc kubenswrapper[4817]: E1128 15:33:33.777526 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:33:37 crc kubenswrapper[4817]: I1128 15:33:37.716271 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_789d2be2-8615-428f-bc54-68d019823a99/memcached/0.log" Nov 28 15:33:48 crc kubenswrapper[4817]: I1128 15:33:48.737122 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:33:48 crc kubenswrapper[4817]: E1128 15:33:48.737745 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:34:00 crc kubenswrapper[4817]: I1128 15:34:00.736414 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:34:00 crc kubenswrapper[4817]: E1128 15:34:00.737124 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:34:00 crc kubenswrapper[4817]: I1128 15:34:00.958131 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-8lb8j_59388928-9001-4abd-9361-d4045b4d027b/kube-rbac-proxy/0.log" Nov 28 15:34:00 crc kubenswrapper[4817]: I1128 15:34:00.963769 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-8lb8j_59388928-9001-4abd-9361-d4045b4d027b/manager/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.177132 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-75m4b_b0544c31-6ad6-4a76-84da-c180f89ba5ee/kube-rbac-proxy/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.272908 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-75m4b_b0544c31-6ad6-4a76-84da-c180f89ba5ee/manager/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.353380 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/util/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.494189 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/pull/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.504020 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/util/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.537151 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/pull/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.700259 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/util/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.706394 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/extract/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.718361 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/pull/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.890616 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-8cjxb_025353e3-a121-42b5-b283-727b6eb24b7c/kube-rbac-proxy/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.893746 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-8cjxb_025353e3-a121-42b5-b283-727b6eb24b7c/manager/0.log" Nov 28 15:34:01 crc kubenswrapper[4817]: I1128 15:34:01.989900 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-4fj4z_29cc111e-478d-48c6-a89c-4050114b5d37/kube-rbac-proxy/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.138392 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-4fj4z_29cc111e-478d-48c6-a89c-4050114b5d37/manager/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.174926 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-k7tbk_02f94412-5d1b-401d-a9fa-211df97a2710/kube-rbac-proxy/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.254057 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-k7tbk_02f94412-5d1b-401d-a9fa-211df97a2710/manager/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.328341 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-j2vsj_e04c0207-7d18-437a-81a3-51ca263e492b/kube-rbac-proxy/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.351816 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-j2vsj_e04c0207-7d18-437a-81a3-51ca263e492b/manager/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.531587 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-w9qqs_a55aea0f-52dc-4403-8067-2372b24cdf83/kube-rbac-proxy/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.683218 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-w9qqs_a55aea0f-52dc-4403-8067-2372b24cdf83/manager/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.696062 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-vh4rl_4598ac96-dead-47cc-b063-22d3197689a4/kube-rbac-proxy/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.734712 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-vh4rl_4598ac96-dead-47cc-b063-22d3197689a4/manager/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.845239 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4phnt_694e1cb6-aa3a-4909-b220-99d42dbf1e49/kube-rbac-proxy/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.970134 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4phnt_694e1cb6-aa3a-4909-b220-99d42dbf1e49/manager/0.log" Nov 28 15:34:02 crc kubenswrapper[4817]: I1128 15:34:02.992010 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-tbssr_be902d34-5e5e-48be-9a6b-d9e53661f367/kube-rbac-proxy/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.068620 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-tbssr_be902d34-5e5e-48be-9a6b-d9e53661f367/manager/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.138644 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-87bqn_19220366-8f0a-4108-9093-964f54a4642f/kube-rbac-proxy/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.227497 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-87bqn_19220366-8f0a-4108-9093-964f54a4642f/manager/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.302263 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-pmw8p_2f81ca1e-831f-4a15-9ea9-a81364b534fd/kube-rbac-proxy/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.388611 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-pmw8p_2f81ca1e-831f-4a15-9ea9-a81364b534fd/manager/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.477026 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-wzw2n_abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29/kube-rbac-proxy/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.579028 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-wzw2n_abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29/manager/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.629905 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-gv25c_18412c56-76ce-48a3-81f2-3cbc4d090df5/kube-rbac-proxy/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.685443 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-gv25c_18412c56-76ce-48a3-81f2-3cbc4d090df5/manager/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.808995 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm_35934979-2374-4a86-b9db-7effb7624f40/kube-rbac-proxy/0.log" Nov 28 15:34:03 crc kubenswrapper[4817]: I1128 15:34:03.875195 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm_35934979-2374-4a86-b9db-7effb7624f40/manager/0.log" Nov 28 15:34:04 crc kubenswrapper[4817]: I1128 15:34:04.271911 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56ffd7cc6d-rw9vh_cb0ec512-1d2d-4487-8d7e-c412ea2b4da6/operator/0.log" Nov 28 15:34:04 crc kubenswrapper[4817]: I1128 15:34:04.332033 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mt7jk_3efb685f-68cf-4455-9e24-ac2d54d45dab/registry-server/0.log" Nov 28 15:34:04 crc kubenswrapper[4817]: I1128 15:34:04.418894 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-s52pg_6ec7b27a-79e7-402c-8b73-091c9c20a1db/kube-rbac-proxy/0.log" Nov 28 15:34:04 crc kubenswrapper[4817]: I1128 15:34:04.528572 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-s52pg_6ec7b27a-79e7-402c-8b73-091c9c20a1db/manager/0.log" Nov 28 15:34:04 crc kubenswrapper[4817]: I1128 15:34:04.576271 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8fqs4_5343b3a1-d2c2-4b26-806b-a7661275d20f/kube-rbac-proxy/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.483308 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8fqs4_5343b3a1-d2c2-4b26-806b-a7661275d20f/manager/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.548494 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jr75r_d9e5f208-4552-48ff-af77-901fa7c76e42/operator/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.610998 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64d6f76c7b-5smh2_c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5/manager/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.692352 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-n96rm_a023b80a-0d1c-47c6-b904-014d659e559a/kube-rbac-proxy/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.699205 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-n96rm_a023b80a-0d1c-47c6-b904-014d659e559a/manager/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.748412 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/kube-rbac-proxy/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.908982 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-xpvwd_e8b4fc65-1b95-41fb-a95e-6e102c452691/kube-rbac-proxy/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.955403 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:34:05 crc kubenswrapper[4817]: I1128 15:34:05.961900 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-xpvwd_e8b4fc65-1b95-41fb-a95e-6e102c452691/manager/0.log" Nov 28 15:34:06 crc kubenswrapper[4817]: I1128 15:34:06.113577 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-26v9h_77766fad-8504-4053-a3f6-78591e568ce5/kube-rbac-proxy/0.log" Nov 28 15:34:06 crc kubenswrapper[4817]: I1128 15:34:06.114810 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-26v9h_77766fad-8504-4053-a3f6-78591e568ce5/manager/0.log" Nov 28 15:34:12 crc kubenswrapper[4817]: I1128 15:34:12.737856 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:34:12 crc kubenswrapper[4817]: E1128 15:34:12.738626 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:34:25 crc kubenswrapper[4817]: I1128 15:34:25.738116 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:34:25 crc kubenswrapper[4817]: E1128 15:34:25.739265 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:34:26 crc kubenswrapper[4817]: I1128 15:34:26.429095 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-kvssf_55dbc750-3c63-44f9-ada1-aa282adb3562/control-plane-machine-set-operator/0.log" Nov 28 15:34:26 crc kubenswrapper[4817]: I1128 15:34:26.620298 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-skq7g_e20cea1b-1075-4334-a1b6-3634ee774393/kube-rbac-proxy/0.log" Nov 28 15:34:26 crc kubenswrapper[4817]: I1128 15:34:26.690388 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-skq7g_e20cea1b-1075-4334-a1b6-3634ee774393/machine-api-operator/0.log" Nov 28 15:34:38 crc kubenswrapper[4817]: I1128 15:34:38.737343 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:34:38 crc kubenswrapper[4817]: E1128 15:34:38.737995 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:34:39 crc kubenswrapper[4817]: I1128 15:34:39.146096 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-d2rzq_d33ac74c-861f-4b72-8ba8-38e14244b510/cert-manager-controller/0.log" Nov 28 15:34:39 crc kubenswrapper[4817]: I1128 15:34:39.217934 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-fmdv9_d6857358-29dc-4bfa-89cc-503c7d452236/cert-manager-cainjector/0.log" Nov 28 15:34:39 crc kubenswrapper[4817]: I1128 15:34:39.308592 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xfd76_aee642f7-2003-406e-b5fd-b5a9b6c6e1fe/cert-manager-webhook/0.log" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.650791 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-25plv"] Nov 28 15:34:43 crc kubenswrapper[4817]: E1128 15:34:43.651797 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="extract-utilities" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.651811 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="extract-utilities" Nov 28 15:34:43 crc kubenswrapper[4817]: E1128 15:34:43.651825 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b91a795-5e03-4b0f-9eba-2a57a0c38078" containerName="container-00" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.651830 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b91a795-5e03-4b0f-9eba-2a57a0c38078" containerName="container-00" Nov 28 15:34:43 crc kubenswrapper[4817]: E1128 15:34:43.651865 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="extract-content" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.651871 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="extract-content" Nov 28 15:34:43 crc kubenswrapper[4817]: E1128 15:34:43.651883 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="registry-server" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.651889 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="registry-server" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.652075 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7f71ec-3b13-40be-9c77-bf1783abfb80" containerName="registry-server" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.652088 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b91a795-5e03-4b0f-9eba-2a57a0c38078" containerName="container-00" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.653457 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.679031 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25plv"] Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.840104 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-utilities\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.840667 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-catalog-content\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.840775 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfbpd\" (UniqueName: \"kubernetes.io/projected/591e67c7-c7f2-426c-87de-64848137b943-kube-api-access-hfbpd\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.943495 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-utilities\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.943593 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-catalog-content\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.943630 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfbpd\" (UniqueName: \"kubernetes.io/projected/591e67c7-c7f2-426c-87de-64848137b943-kube-api-access-hfbpd\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.944101 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-utilities\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.944270 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-catalog-content\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.969912 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfbpd\" (UniqueName: \"kubernetes.io/projected/591e67c7-c7f2-426c-87de-64848137b943-kube-api-access-hfbpd\") pod \"redhat-marketplace-25plv\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:43 crc kubenswrapper[4817]: I1128 15:34:43.974408 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:44 crc kubenswrapper[4817]: I1128 15:34:44.481249 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25plv"] Nov 28 15:34:45 crc kubenswrapper[4817]: I1128 15:34:45.511930 4817 generic.go:334] "Generic (PLEG): container finished" podID="591e67c7-c7f2-426c-87de-64848137b943" containerID="e29b3cf1da4c910e358cc1ab6d36694b10ca3149aecc00702a1372bca24f2421" exitCode=0 Nov 28 15:34:45 crc kubenswrapper[4817]: I1128 15:34:45.512021 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerDied","Data":"e29b3cf1da4c910e358cc1ab6d36694b10ca3149aecc00702a1372bca24f2421"} Nov 28 15:34:45 crc kubenswrapper[4817]: I1128 15:34:45.512472 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerStarted","Data":"34e53219699befece0719c9fa9241f0cbf287d774990095866e32a6c8df26e83"} Nov 28 15:34:46 crc kubenswrapper[4817]: I1128 15:34:46.523357 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerStarted","Data":"7e465a3c3812338133be9e86fcbc415dbabab4178ac17a5bc0a6bf8ac12e95bd"} Nov 28 15:34:47 crc kubenswrapper[4817]: I1128 15:34:47.541653 4817 generic.go:334] "Generic (PLEG): container finished" podID="591e67c7-c7f2-426c-87de-64848137b943" containerID="7e465a3c3812338133be9e86fcbc415dbabab4178ac17a5bc0a6bf8ac12e95bd" exitCode=0 Nov 28 15:34:47 crc kubenswrapper[4817]: I1128 15:34:47.541883 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerDied","Data":"7e465a3c3812338133be9e86fcbc415dbabab4178ac17a5bc0a6bf8ac12e95bd"} Nov 28 15:34:48 crc kubenswrapper[4817]: I1128 15:34:48.558245 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerStarted","Data":"fb19faef410c997c9af523794eae83058983f7b549cc3f2e4279971279b47e27"} Nov 28 15:34:48 crc kubenswrapper[4817]: I1128 15:34:48.591540 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-25plv" podStartSLOduration=3.074467041 podStartE2EDuration="5.59151463s" podCreationTimestamp="2025-11-28 15:34:43 +0000 UTC" firstStartedPulling="2025-11-28 15:34:45.513875473 +0000 UTC m=+3908.101853739" lastFinishedPulling="2025-11-28 15:34:48.030923052 +0000 UTC m=+3910.618901328" observedRunningTime="2025-11-28 15:34:48.579837748 +0000 UTC m=+3911.167816024" watchObservedRunningTime="2025-11-28 15:34:48.59151463 +0000 UTC m=+3911.179492936" Nov 28 15:34:49 crc kubenswrapper[4817]: I1128 15:34:49.736563 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:34:49 crc kubenswrapper[4817]: E1128 15:34:49.737183 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.436823 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-hzvx2_f9dc9469-cf27-4d53-ae6d-a4df72d8af78/nmstate-console-plugin/0.log" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.565084 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kgbh6_feed3d1f-a072-4ccc-b47a-3a5b48b44177/nmstate-handler/0.log" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.627905 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-wgrqq_90fcff6a-9f08-479c-8e5b-2745700bc8f1/kube-rbac-proxy/0.log" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.691051 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-wgrqq_90fcff6a-9f08-479c-8e5b-2745700bc8f1/nmstate-metrics/0.log" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.822981 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-b4g4f_68d8ec1c-4246-402e-8da4-53dfe24edbca/nmstate-operator/0.log" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.897072 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-pm78n_8d2164f8-732d-4650-b92a-3fb677b51c25/nmstate-webhook/0.log" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.975493 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:53 crc kubenswrapper[4817]: I1128 15:34:53.975544 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:54 crc kubenswrapper[4817]: I1128 15:34:54.029036 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:54 crc kubenswrapper[4817]: I1128 15:34:54.671039 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:54 crc kubenswrapper[4817]: I1128 15:34:54.721639 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25plv"] Nov 28 15:34:56 crc kubenswrapper[4817]: I1128 15:34:56.634190 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-25plv" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="registry-server" containerID="cri-o://fb19faef410c997c9af523794eae83058983f7b549cc3f2e4279971279b47e27" gracePeriod=2 Nov 28 15:34:57 crc kubenswrapper[4817]: I1128 15:34:57.644100 4817 generic.go:334] "Generic (PLEG): container finished" podID="591e67c7-c7f2-426c-87de-64848137b943" containerID="fb19faef410c997c9af523794eae83058983f7b549cc3f2e4279971279b47e27" exitCode=0 Nov 28 15:34:57 crc kubenswrapper[4817]: I1128 15:34:57.644294 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerDied","Data":"fb19faef410c997c9af523794eae83058983f7b549cc3f2e4279971279b47e27"} Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.472682 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.522777 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-catalog-content\") pod \"591e67c7-c7f2-426c-87de-64848137b943\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.522865 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfbpd\" (UniqueName: \"kubernetes.io/projected/591e67c7-c7f2-426c-87de-64848137b943-kube-api-access-hfbpd\") pod \"591e67c7-c7f2-426c-87de-64848137b943\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.523087 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-utilities\") pod \"591e67c7-c7f2-426c-87de-64848137b943\" (UID: \"591e67c7-c7f2-426c-87de-64848137b943\") " Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.524284 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-utilities" (OuterVolumeSpecName: "utilities") pod "591e67c7-c7f2-426c-87de-64848137b943" (UID: "591e67c7-c7f2-426c-87de-64848137b943"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.534093 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/591e67c7-c7f2-426c-87de-64848137b943-kube-api-access-hfbpd" (OuterVolumeSpecName: "kube-api-access-hfbpd") pod "591e67c7-c7f2-426c-87de-64848137b943" (UID: "591e67c7-c7f2-426c-87de-64848137b943"). InnerVolumeSpecName "kube-api-access-hfbpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.544331 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "591e67c7-c7f2-426c-87de-64848137b943" (UID: "591e67c7-c7f2-426c-87de-64848137b943"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.625240 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.625279 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591e67c7-c7f2-426c-87de-64848137b943-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.625291 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfbpd\" (UniqueName: \"kubernetes.io/projected/591e67c7-c7f2-426c-87de-64848137b943-kube-api-access-hfbpd\") on node \"crc\" DevicePath \"\"" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.656566 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25plv" event={"ID":"591e67c7-c7f2-426c-87de-64848137b943","Type":"ContainerDied","Data":"34e53219699befece0719c9fa9241f0cbf287d774990095866e32a6c8df26e83"} Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.656631 4817 scope.go:117] "RemoveContainer" containerID="fb19faef410c997c9af523794eae83058983f7b549cc3f2e4279971279b47e27" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.656664 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25plv" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.690914 4817 scope.go:117] "RemoveContainer" containerID="7e465a3c3812338133be9e86fcbc415dbabab4178ac17a5bc0a6bf8ac12e95bd" Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.711793 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25plv"] Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.721506 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-25plv"] Nov 28 15:34:58 crc kubenswrapper[4817]: I1128 15:34:58.724923 4817 scope.go:117] "RemoveContainer" containerID="e29b3cf1da4c910e358cc1ab6d36694b10ca3149aecc00702a1372bca24f2421" Nov 28 15:34:59 crc kubenswrapper[4817]: I1128 15:34:59.747822 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="591e67c7-c7f2-426c-87de-64848137b943" path="/var/lib/kubelet/pods/591e67c7-c7f2-426c-87de-64848137b943/volumes" Nov 28 15:35:03 crc kubenswrapper[4817]: I1128 15:35:03.737905 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:35:03 crc kubenswrapper[4817]: E1128 15:35:03.739301 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.296754 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-4wsbn_7f4b2cee-aa3d-43ca-adbd-42dee238c7bc/kube-rbac-proxy/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.384288 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-4wsbn_7f4b2cee-aa3d-43ca-adbd-42dee238c7bc/controller/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.503007 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.676360 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.697421 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.724572 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.770918 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.927840 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.928122 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:35:09 crc kubenswrapper[4817]: I1128 15:35:09.936065 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.002454 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.178133 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.209467 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.211835 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.256301 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/controller/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.424429 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/frr-metrics/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.424528 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/kube-rbac-proxy/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.493949 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/kube-rbac-proxy-frr/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.689419 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/reloader/0.log" Nov 28 15:35:10 crc kubenswrapper[4817]: I1128 15:35:10.758279 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-4r56x_8f400971-fd8a-4c48-82bd-33f47438bcb0/frr-k8s-webhook-server/0.log" Nov 28 15:35:11 crc kubenswrapper[4817]: I1128 15:35:11.039496 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d674c8684-kk2kj_aebf367a-7ec1-4675-b01f-3d54d2ba0613/manager/0.log" Nov 28 15:35:11 crc kubenswrapper[4817]: I1128 15:35:11.272695 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74f997fcc-g6nm4_5980acb2-a5d3-44fa-8bf0-23e36cde74a4/webhook-server/0.log" Nov 28 15:35:11 crc kubenswrapper[4817]: I1128 15:35:11.302506 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkbsn_f737005e-fcc0-4f24-bdbb-99d7f22afef6/kube-rbac-proxy/0.log" Nov 28 15:35:11 crc kubenswrapper[4817]: I1128 15:35:11.989713 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkbsn_f737005e-fcc0-4f24-bdbb-99d7f22afef6/speaker/0.log" Nov 28 15:35:12 crc kubenswrapper[4817]: I1128 15:35:12.149425 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/frr/0.log" Nov 28 15:35:17 crc kubenswrapper[4817]: I1128 15:35:17.743707 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:35:17 crc kubenswrapper[4817]: E1128 15:35:17.744554 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.552879 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/util/0.log" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.731127 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/pull/0.log" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.758062 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/util/0.log" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.797278 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/pull/0.log" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.964467 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/extract/0.log" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.973382 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/util/0.log" Nov 28 15:35:25 crc kubenswrapper[4817]: I1128 15:35:25.986995 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/pull/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.133772 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/util/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.324489 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/util/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.336501 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/pull/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.349702 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/pull/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.530399 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/util/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.539822 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/pull/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.553083 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/extract/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.717939 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/util/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.896282 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/pull/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.907322 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/util/0.log" Nov 28 15:35:26 crc kubenswrapper[4817]: I1128 15:35:26.918092 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/pull/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.118704 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/util/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.126651 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/pull/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.194089 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/extract/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.339449 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-utilities/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.511351 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-content/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.530639 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-utilities/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.551178 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-content/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.692602 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-utilities/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.743316 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-content/0.log" Nov 28 15:35:27 crc kubenswrapper[4817]: I1128 15:35:27.929746 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-utilities/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.209061 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-content/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.223671 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-content/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.271224 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/registry-server/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.297174 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-utilities/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.497882 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-content/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.543181 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-utilities/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.844371 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hxjk5_9482d966-5b89-4687-9b4a-29eefc683dcd/marketplace-operator/0.log" Nov 28 15:35:28 crc kubenswrapper[4817]: I1128 15:35:28.993227 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-utilities/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.118932 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/registry-server/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.222754 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-utilities/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.223910 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-content/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.230846 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-content/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.409074 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-utilities/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.470841 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-content/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.534554 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/registry-server/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.544438 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-utilities/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.697496 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-content/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.710709 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-content/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.722454 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-utilities/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.897369 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-utilities/0.log" Nov 28 15:35:29 crc kubenswrapper[4817]: I1128 15:35:29.927879 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-content/0.log" Nov 28 15:35:30 crc kubenswrapper[4817]: I1128 15:35:30.368214 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/registry-server/0.log" Nov 28 15:35:30 crc kubenswrapper[4817]: I1128 15:35:30.736680 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:35:30 crc kubenswrapper[4817]: E1128 15:35:30.737060 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:35:42 crc kubenswrapper[4817]: I1128 15:35:42.737364 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:35:42 crc kubenswrapper[4817]: E1128 15:35:42.738210 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:35:43 crc kubenswrapper[4817]: I1128 15:35:43.113807 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-jvrtq_ad4b53c3-5d16-412b-ba96-f4ba1a54624a/prometheus-operator/0.log" Nov 28 15:35:43 crc kubenswrapper[4817]: I1128 15:35:43.247022 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw_4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf/prometheus-operator-admission-webhook/0.log" Nov 28 15:35:43 crc kubenswrapper[4817]: I1128 15:35:43.371808 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm_597ad435-7283-469d-8c9a-b0e9a032a433/prometheus-operator-admission-webhook/0.log" Nov 28 15:35:43 crc kubenswrapper[4817]: I1128 15:35:43.498372 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-7mc44_9bd8cd77-0f23-40cc-87aa-19ef553d9565/operator/0.log" Nov 28 15:35:43 crc kubenswrapper[4817]: I1128 15:35:43.584692 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-gcb2c_334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f/perses-operator/0.log" Nov 28 15:35:56 crc kubenswrapper[4817]: I1128 15:35:56.739699 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:35:56 crc kubenswrapper[4817]: E1128 15:35:56.740772 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:36:01 crc kubenswrapper[4817]: E1128 15:36:01.064673 4817 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.236:34876->38.102.83.236:36927: write tcp 38.102.83.236:34876->38.102.83.236:36927: write: broken pipe Nov 28 15:36:08 crc kubenswrapper[4817]: I1128 15:36:08.737856 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:36:08 crc kubenswrapper[4817]: E1128 15:36:08.738825 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:36:21 crc kubenswrapper[4817]: I1128 15:36:21.738369 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:36:21 crc kubenswrapper[4817]: E1128 15:36:21.739371 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:36:34 crc kubenswrapper[4817]: I1128 15:36:34.738555 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:36:35 crc kubenswrapper[4817]: I1128 15:36:35.755306 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"898fa3fe56fad6eb55fca0346a27b58f52616da04be43eac8a66a01390e66f20"} Nov 28 15:37:09 crc kubenswrapper[4817]: I1128 15:37:09.126380 4817 generic.go:334] "Generic (PLEG): container finished" podID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerID="e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68" exitCode=0 Nov 28 15:37:09 crc kubenswrapper[4817]: I1128 15:37:09.126505 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" event={"ID":"ebf5c67c-81e6-4878-a23c-9f45e14f6697","Type":"ContainerDied","Data":"e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68"} Nov 28 15:37:09 crc kubenswrapper[4817]: I1128 15:37:09.127410 4817 scope.go:117] "RemoveContainer" containerID="e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68" Nov 28 15:37:09 crc kubenswrapper[4817]: I1128 15:37:09.882663 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fxmv8_must-gather-tdmd6_ebf5c67c-81e6-4878-a23c-9f45e14f6697/gather/0.log" Nov 28 15:37:17 crc kubenswrapper[4817]: I1128 15:37:17.570967 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fxmv8/must-gather-tdmd6"] Nov 28 15:37:17 crc kubenswrapper[4817]: I1128 15:37:17.571883 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="copy" containerID="cri-o://94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6" gracePeriod=2 Nov 28 15:37:17 crc kubenswrapper[4817]: I1128 15:37:17.578914 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fxmv8/must-gather-tdmd6"] Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.111303 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fxmv8_must-gather-tdmd6_ebf5c67c-81e6-4878-a23c-9f45e14f6697/copy/0.log" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.112006 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.224100 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fxmv8_must-gather-tdmd6_ebf5c67c-81e6-4878-a23c-9f45e14f6697/copy/0.log" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.224513 4817 generic.go:334] "Generic (PLEG): container finished" podID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerID="94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6" exitCode=143 Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.224578 4817 scope.go:117] "RemoveContainer" containerID="94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.224594 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fxmv8/must-gather-tdmd6" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.232897 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ebf5c67c-81e6-4878-a23c-9f45e14f6697-must-gather-output\") pod \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.232958 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr5gw\" (UniqueName: \"kubernetes.io/projected/ebf5c67c-81e6-4878-a23c-9f45e14f6697-kube-api-access-wr5gw\") pod \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\" (UID: \"ebf5c67c-81e6-4878-a23c-9f45e14f6697\") " Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.240112 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf5c67c-81e6-4878-a23c-9f45e14f6697-kube-api-access-wr5gw" (OuterVolumeSpecName: "kube-api-access-wr5gw") pod "ebf5c67c-81e6-4878-a23c-9f45e14f6697" (UID: "ebf5c67c-81e6-4878-a23c-9f45e14f6697"). InnerVolumeSpecName "kube-api-access-wr5gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.256454 4817 scope.go:117] "RemoveContainer" containerID="e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.336278 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr5gw\" (UniqueName: \"kubernetes.io/projected/ebf5c67c-81e6-4878-a23c-9f45e14f6697-kube-api-access-wr5gw\") on node \"crc\" DevicePath \"\"" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.375579 4817 scope.go:117] "RemoveContainer" containerID="94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6" Nov 28 15:37:18 crc kubenswrapper[4817]: E1128 15:37:18.376096 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6\": container with ID starting with 94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6 not found: ID does not exist" containerID="94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.376129 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6"} err="failed to get container status \"94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6\": rpc error: code = NotFound desc = could not find container \"94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6\": container with ID starting with 94ac971939945f18b43ca83a7e1ea14f71641ad5387707a8feb596b9aa3f53d6 not found: ID does not exist" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.376147 4817 scope.go:117] "RemoveContainer" containerID="e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68" Nov 28 15:37:18 crc kubenswrapper[4817]: E1128 15:37:18.376407 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68\": container with ID starting with e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68 not found: ID does not exist" containerID="e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.376430 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68"} err="failed to get container status \"e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68\": rpc error: code = NotFound desc = could not find container \"e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68\": container with ID starting with e9b1be586c8a0fdb1c58b7bd37959f6656437494b773d812ca22ec33bcadda68 not found: ID does not exist" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.394393 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf5c67c-81e6-4878-a23c-9f45e14f6697-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ebf5c67c-81e6-4878-a23c-9f45e14f6697" (UID: "ebf5c67c-81e6-4878-a23c-9f45e14f6697"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:37:18 crc kubenswrapper[4817]: I1128 15:37:18.437809 4817 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ebf5c67c-81e6-4878-a23c-9f45e14f6697-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 15:37:19 crc kubenswrapper[4817]: I1128 15:37:19.747073 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" path="/var/lib/kubelet/pods/ebf5c67c-81e6-4878-a23c-9f45e14f6697/volumes" Nov 28 15:38:15 crc kubenswrapper[4817]: I1128 15:38:15.784461 4817 scope.go:117] "RemoveContainer" containerID="25f50de32384def93bf3c1878ae2ad10e365233d731c5c3ac129c58c0d27d097" Nov 28 15:38:17 crc kubenswrapper[4817]: I1128 15:38:17.030097 4817 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-87bqn" podUID="19220366-8f0a-4108-9093-964f54a4642f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 15:38:54 crc kubenswrapper[4817]: I1128 15:38:54.045634 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:38:54 crc kubenswrapper[4817]: I1128 15:38:54.046250 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:39:24 crc kubenswrapper[4817]: I1128 15:39:24.045538 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:39:24 crc kubenswrapper[4817]: I1128 15:39:24.046166 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.366852 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6ff45"] Nov 28 15:39:40 crc kubenswrapper[4817]: E1128 15:39:40.368374 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="copy" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.368403 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="copy" Nov 28 15:39:40 crc kubenswrapper[4817]: E1128 15:39:40.368423 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="extract-content" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.368435 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="extract-content" Nov 28 15:39:40 crc kubenswrapper[4817]: E1128 15:39:40.368477 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="registry-server" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.368490 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="registry-server" Nov 28 15:39:40 crc kubenswrapper[4817]: E1128 15:39:40.368517 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="extract-utilities" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.368531 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="extract-utilities" Nov 28 15:39:40 crc kubenswrapper[4817]: E1128 15:39:40.368551 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="gather" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.368567 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="gather" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.369076 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="gather" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.369131 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="591e67c7-c7f2-426c-87de-64848137b943" containerName="registry-server" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.369162 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf5c67c-81e6-4878-a23c-9f45e14f6697" containerName="copy" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.374098 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.387605 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ff45"] Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.505904 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-catalog-content\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.506354 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-utilities\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.506428 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsphm\" (UniqueName: \"kubernetes.io/projected/ea951c65-882a-4484-9397-09c5e69131b9-kube-api-access-vsphm\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.608054 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-utilities\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.608130 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsphm\" (UniqueName: \"kubernetes.io/projected/ea951c65-882a-4484-9397-09c5e69131b9-kube-api-access-vsphm\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.608211 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-catalog-content\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.608694 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-utilities\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.608740 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-catalog-content\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.629676 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsphm\" (UniqueName: \"kubernetes.io/projected/ea951c65-882a-4484-9397-09c5e69131b9-kube-api-access-vsphm\") pod \"certified-operators-6ff45\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:40 crc kubenswrapper[4817]: I1128 15:39:40.698606 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:41 crc kubenswrapper[4817]: I1128 15:39:41.211111 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ff45"] Nov 28 15:39:41 crc kubenswrapper[4817]: I1128 15:39:41.732927 4817 generic.go:334] "Generic (PLEG): container finished" podID="ea951c65-882a-4484-9397-09c5e69131b9" containerID="8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511" exitCode=0 Nov 28 15:39:41 crc kubenswrapper[4817]: I1128 15:39:41.732990 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerDied","Data":"8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511"} Nov 28 15:39:41 crc kubenswrapper[4817]: I1128 15:39:41.733249 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerStarted","Data":"a250f5d0b9e7b79e9d8b03f2da0af25b0d6a62937ff5aeae930a1c6288583015"} Nov 28 15:39:41 crc kubenswrapper[4817]: I1128 15:39:41.736113 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:39:43 crc kubenswrapper[4817]: I1128 15:39:43.761091 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerStarted","Data":"256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c"} Nov 28 15:39:44 crc kubenswrapper[4817]: I1128 15:39:44.776246 4817 generic.go:334] "Generic (PLEG): container finished" podID="ea951c65-882a-4484-9397-09c5e69131b9" containerID="256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c" exitCode=0 Nov 28 15:39:44 crc kubenswrapper[4817]: I1128 15:39:44.776281 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerDied","Data":"256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c"} Nov 28 15:39:46 crc kubenswrapper[4817]: I1128 15:39:46.796488 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerStarted","Data":"25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f"} Nov 28 15:39:46 crc kubenswrapper[4817]: I1128 15:39:46.825209 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6ff45" podStartSLOduration=2.7548678300000002 podStartE2EDuration="6.825188346s" podCreationTimestamp="2025-11-28 15:39:40 +0000 UTC" firstStartedPulling="2025-11-28 15:39:41.735815647 +0000 UTC m=+4204.323793913" lastFinishedPulling="2025-11-28 15:39:45.806136163 +0000 UTC m=+4208.394114429" observedRunningTime="2025-11-28 15:39:46.820102212 +0000 UTC m=+4209.408080498" watchObservedRunningTime="2025-11-28 15:39:46.825188346 +0000 UTC m=+4209.413166622" Nov 28 15:39:50 crc kubenswrapper[4817]: I1128 15:39:50.699107 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:50 crc kubenswrapper[4817]: I1128 15:39:50.699676 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:50 crc kubenswrapper[4817]: I1128 15:39:50.768748 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:50 crc kubenswrapper[4817]: I1128 15:39:50.917573 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:51 crc kubenswrapper[4817]: I1128 15:39:51.012503 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ff45"] Nov 28 15:39:52 crc kubenswrapper[4817]: I1128 15:39:52.860250 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6ff45" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="registry-server" containerID="cri-o://25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f" gracePeriod=2 Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.847853 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.879575 4817 generic.go:334] "Generic (PLEG): container finished" podID="ea951c65-882a-4484-9397-09c5e69131b9" containerID="25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f" exitCode=0 Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.879651 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerDied","Data":"25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f"} Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.879706 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ff45" event={"ID":"ea951c65-882a-4484-9397-09c5e69131b9","Type":"ContainerDied","Data":"a250f5d0b9e7b79e9d8b03f2da0af25b0d6a62937ff5aeae930a1c6288583015"} Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.879738 4817 scope.go:117] "RemoveContainer" containerID="25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.879695 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ff45" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.900487 4817 scope.go:117] "RemoveContainer" containerID="256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.927885 4817 scope.go:117] "RemoveContainer" containerID="8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.965690 4817 scope.go:117] "RemoveContainer" containerID="25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f" Nov 28 15:39:53 crc kubenswrapper[4817]: E1128 15:39:53.966145 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f\": container with ID starting with 25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f not found: ID does not exist" containerID="25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.966179 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f"} err="failed to get container status \"25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f\": rpc error: code = NotFound desc = could not find container \"25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f\": container with ID starting with 25b7d48d180679056b21c0e02a9e71fd5fedc62a6a91a78e3693210824222c4f not found: ID does not exist" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.966199 4817 scope.go:117] "RemoveContainer" containerID="256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c" Nov 28 15:39:53 crc kubenswrapper[4817]: E1128 15:39:53.966469 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c\": container with ID starting with 256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c not found: ID does not exist" containerID="256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.966554 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c"} err="failed to get container status \"256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c\": rpc error: code = NotFound desc = could not find container \"256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c\": container with ID starting with 256c4b1954f07d2f7b2a5b24271deaefeae0aa3c33b0977637632f7a4869530c not found: ID does not exist" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.966585 4817 scope.go:117] "RemoveContainer" containerID="8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511" Nov 28 15:39:53 crc kubenswrapper[4817]: E1128 15:39:53.966932 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511\": container with ID starting with 8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511 not found: ID does not exist" containerID="8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511" Nov 28 15:39:53 crc kubenswrapper[4817]: I1128 15:39:53.966997 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511"} err="failed to get container status \"8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511\": rpc error: code = NotFound desc = could not find container \"8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511\": container with ID starting with 8e2614f312a43acc2daa9b8677f60b498697f1afec1f7930bb5cc35dc786c511 not found: ID does not exist" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.003047 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsphm\" (UniqueName: \"kubernetes.io/projected/ea951c65-882a-4484-9397-09c5e69131b9-kube-api-access-vsphm\") pod \"ea951c65-882a-4484-9397-09c5e69131b9\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.003237 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-catalog-content\") pod \"ea951c65-882a-4484-9397-09c5e69131b9\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.003322 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-utilities\") pod \"ea951c65-882a-4484-9397-09c5e69131b9\" (UID: \"ea951c65-882a-4484-9397-09c5e69131b9\") " Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.004621 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-utilities" (OuterVolumeSpecName: "utilities") pod "ea951c65-882a-4484-9397-09c5e69131b9" (UID: "ea951c65-882a-4484-9397-09c5e69131b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.011275 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea951c65-882a-4484-9397-09c5e69131b9-kube-api-access-vsphm" (OuterVolumeSpecName: "kube-api-access-vsphm") pod "ea951c65-882a-4484-9397-09c5e69131b9" (UID: "ea951c65-882a-4484-9397-09c5e69131b9"). InnerVolumeSpecName "kube-api-access-vsphm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.045529 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.045629 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.045670 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.046428 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"898fa3fe56fad6eb55fca0346a27b58f52616da04be43eac8a66a01390e66f20"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.046485 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://898fa3fe56fad6eb55fca0346a27b58f52616da04be43eac8a66a01390e66f20" gracePeriod=600 Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.059884 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea951c65-882a-4484-9397-09c5e69131b9" (UID: "ea951c65-882a-4484-9397-09c5e69131b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.105109 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.105146 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsphm\" (UniqueName: \"kubernetes.io/projected/ea951c65-882a-4484-9397-09c5e69131b9-kube-api-access-vsphm\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.105159 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea951c65-882a-4484-9397-09c5e69131b9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.233010 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ff45"] Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.247091 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6ff45"] Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.891636 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="898fa3fe56fad6eb55fca0346a27b58f52616da04be43eac8a66a01390e66f20" exitCode=0 Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.891738 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"898fa3fe56fad6eb55fca0346a27b58f52616da04be43eac8a66a01390e66f20"} Nov 28 15:39:54 crc kubenswrapper[4817]: I1128 15:39:54.891785 4817 scope.go:117] "RemoveContainer" containerID="59511d0b5bd63ca3dc55dfa33a99089073376b781859883772f30c129b73a068" Nov 28 15:39:55 crc kubenswrapper[4817]: I1128 15:39:55.753307 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea951c65-882a-4484-9397-09c5e69131b9" path="/var/lib/kubelet/pods/ea951c65-882a-4484-9397-09c5e69131b9/volumes" Nov 28 15:39:56 crc kubenswrapper[4817]: I1128 15:39:56.942408 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41"} Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.716985 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9q2hs/must-gather-vcl74"] Nov 28 15:40:18 crc kubenswrapper[4817]: E1128 15:40:18.717887 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="registry-server" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.717899 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="registry-server" Nov 28 15:40:18 crc kubenswrapper[4817]: E1128 15:40:18.717915 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="extract-content" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.717920 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="extract-content" Nov 28 15:40:18 crc kubenswrapper[4817]: E1128 15:40:18.717940 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="extract-utilities" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.717946 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="extract-utilities" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.718170 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea951c65-882a-4484-9397-09c5e69131b9" containerName="registry-server" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.719287 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.726633 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9q2hs"/"default-dockercfg-sthfx" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.726712 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9q2hs"/"openshift-service-ca.crt" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.726922 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9q2hs"/"kube-root-ca.crt" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.740767 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9q2hs/must-gather-vcl74"] Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.791345 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/56977377-a544-492d-803a-a98943ea22fc-must-gather-output\") pod \"must-gather-vcl74\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.791857 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjjz4\" (UniqueName: \"kubernetes.io/projected/56977377-a544-492d-803a-a98943ea22fc-kube-api-access-pjjz4\") pod \"must-gather-vcl74\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.894143 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjjz4\" (UniqueName: \"kubernetes.io/projected/56977377-a544-492d-803a-a98943ea22fc-kube-api-access-pjjz4\") pod \"must-gather-vcl74\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.894248 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/56977377-a544-492d-803a-a98943ea22fc-must-gather-output\") pod \"must-gather-vcl74\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.894783 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/56977377-a544-492d-803a-a98943ea22fc-must-gather-output\") pod \"must-gather-vcl74\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:18 crc kubenswrapper[4817]: I1128 15:40:18.916310 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjjz4\" (UniqueName: \"kubernetes.io/projected/56977377-a544-492d-803a-a98943ea22fc-kube-api-access-pjjz4\") pod \"must-gather-vcl74\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:19 crc kubenswrapper[4817]: I1128 15:40:19.037983 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:40:19 crc kubenswrapper[4817]: I1128 15:40:19.509097 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9q2hs/must-gather-vcl74"] Nov 28 15:40:20 crc kubenswrapper[4817]: I1128 15:40:20.218651 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/must-gather-vcl74" event={"ID":"56977377-a544-492d-803a-a98943ea22fc","Type":"ContainerStarted","Data":"aa6054d5561cd4e3d571a5b17d9e8edc81cfe037dd28257d70118633d5a439c6"} Nov 28 15:40:20 crc kubenswrapper[4817]: I1128 15:40:20.219170 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/must-gather-vcl74" event={"ID":"56977377-a544-492d-803a-a98943ea22fc","Type":"ContainerStarted","Data":"e3cf04c6b2e68290d49e2c14a573ee03261b67aec99fa89dc1c1b92c83b3a72e"} Nov 28 15:40:20 crc kubenswrapper[4817]: I1128 15:40:20.219185 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/must-gather-vcl74" event={"ID":"56977377-a544-492d-803a-a98943ea22fc","Type":"ContainerStarted","Data":"6e3374d4618b4c4dcfae3273f17a752e0ecc0458563a9b2c72bc7a6631569eb0"} Nov 28 15:40:20 crc kubenswrapper[4817]: I1128 15:40:20.266025 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9q2hs/must-gather-vcl74" podStartSLOduration=2.266001557 podStartE2EDuration="2.266001557s" podCreationTimestamp="2025-11-28 15:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:40:20.253123482 +0000 UTC m=+4242.841101758" watchObservedRunningTime="2025-11-28 15:40:20.266001557 +0000 UTC m=+4242.853979823" Nov 28 15:40:23 crc kubenswrapper[4817]: I1128 15:40:23.728673 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9q2hs/crc-debug-br29r"] Nov 28 15:40:23 crc kubenswrapper[4817]: I1128 15:40:23.732766 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:23 crc kubenswrapper[4817]: I1128 15:40:23.914536 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-467xb\" (UniqueName: \"kubernetes.io/projected/1f706cb3-8157-46c4-a297-9eb19562105d-kube-api-access-467xb\") pod \"crc-debug-br29r\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:23 crc kubenswrapper[4817]: I1128 15:40:23.914752 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f706cb3-8157-46c4-a297-9eb19562105d-host\") pod \"crc-debug-br29r\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:24 crc kubenswrapper[4817]: I1128 15:40:24.024885 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-467xb\" (UniqueName: \"kubernetes.io/projected/1f706cb3-8157-46c4-a297-9eb19562105d-kube-api-access-467xb\") pod \"crc-debug-br29r\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:24 crc kubenswrapper[4817]: I1128 15:40:24.025223 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f706cb3-8157-46c4-a297-9eb19562105d-host\") pod \"crc-debug-br29r\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:24 crc kubenswrapper[4817]: I1128 15:40:24.025602 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f706cb3-8157-46c4-a297-9eb19562105d-host\") pod \"crc-debug-br29r\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:24 crc kubenswrapper[4817]: I1128 15:40:24.049091 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-467xb\" (UniqueName: \"kubernetes.io/projected/1f706cb3-8157-46c4-a297-9eb19562105d-kube-api-access-467xb\") pod \"crc-debug-br29r\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:24 crc kubenswrapper[4817]: I1128 15:40:24.064969 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:24 crc kubenswrapper[4817]: W1128 15:40:24.110034 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f706cb3_8157_46c4_a297_9eb19562105d.slice/crio-432abace62843bf5dd6ab0f0e466c1836c8681b581497a8e95165df422f52339 WatchSource:0}: Error finding container 432abace62843bf5dd6ab0f0e466c1836c8681b581497a8e95165df422f52339: Status 404 returned error can't find the container with id 432abace62843bf5dd6ab0f0e466c1836c8681b581497a8e95165df422f52339 Nov 28 15:40:24 crc kubenswrapper[4817]: I1128 15:40:24.262668 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/crc-debug-br29r" event={"ID":"1f706cb3-8157-46c4-a297-9eb19562105d","Type":"ContainerStarted","Data":"432abace62843bf5dd6ab0f0e466c1836c8681b581497a8e95165df422f52339"} Nov 28 15:40:25 crc kubenswrapper[4817]: I1128 15:40:25.275283 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/crc-debug-br29r" event={"ID":"1f706cb3-8157-46c4-a297-9eb19562105d","Type":"ContainerStarted","Data":"2aa5ad1063df5d607952e75f0dd57c3d39b5d5f23112c026bfae0b8cee9666a1"} Nov 28 15:40:25 crc kubenswrapper[4817]: I1128 15:40:25.297763 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9q2hs/crc-debug-br29r" podStartSLOduration=2.297705349 podStartE2EDuration="2.297705349s" podCreationTimestamp="2025-11-28 15:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 15:40:25.29243175 +0000 UTC m=+4247.880410016" watchObservedRunningTime="2025-11-28 15:40:25.297705349 +0000 UTC m=+4247.885683615" Nov 28 15:40:42 crc kubenswrapper[4817]: I1128 15:40:42.447581 4817 generic.go:334] "Generic (PLEG): container finished" podID="1f706cb3-8157-46c4-a297-9eb19562105d" containerID="2aa5ad1063df5d607952e75f0dd57c3d39b5d5f23112c026bfae0b8cee9666a1" exitCode=0 Nov 28 15:40:42 crc kubenswrapper[4817]: I1128 15:40:42.447655 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/crc-debug-br29r" event={"ID":"1f706cb3-8157-46c4-a297-9eb19562105d","Type":"ContainerDied","Data":"2aa5ad1063df5d607952e75f0dd57c3d39b5d5f23112c026bfae0b8cee9666a1"} Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.617555 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.657951 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9q2hs/crc-debug-br29r"] Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.666379 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9q2hs/crc-debug-br29r"] Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.751662 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-467xb\" (UniqueName: \"kubernetes.io/projected/1f706cb3-8157-46c4-a297-9eb19562105d-kube-api-access-467xb\") pod \"1f706cb3-8157-46c4-a297-9eb19562105d\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.751861 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f706cb3-8157-46c4-a297-9eb19562105d-host\") pod \"1f706cb3-8157-46c4-a297-9eb19562105d\" (UID: \"1f706cb3-8157-46c4-a297-9eb19562105d\") " Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.751908 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f706cb3-8157-46c4-a297-9eb19562105d-host" (OuterVolumeSpecName: "host") pod "1f706cb3-8157-46c4-a297-9eb19562105d" (UID: "1f706cb3-8157-46c4-a297-9eb19562105d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.752449 4817 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f706cb3-8157-46c4-a297-9eb19562105d-host\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.770543 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f706cb3-8157-46c4-a297-9eb19562105d-kube-api-access-467xb" (OuterVolumeSpecName: "kube-api-access-467xb") pod "1f706cb3-8157-46c4-a297-9eb19562105d" (UID: "1f706cb3-8157-46c4-a297-9eb19562105d"). InnerVolumeSpecName "kube-api-access-467xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:40:43 crc kubenswrapper[4817]: I1128 15:40:43.854411 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-467xb\" (UniqueName: \"kubernetes.io/projected/1f706cb3-8157-46c4-a297-9eb19562105d-kube-api-access-467xb\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.496896 4817 scope.go:117] "RemoveContainer" containerID="2aa5ad1063df5d607952e75f0dd57c3d39b5d5f23112c026bfae0b8cee9666a1" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.497022 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-br29r" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.842784 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9q2hs/crc-debug-6sfkm"] Nov 28 15:40:44 crc kubenswrapper[4817]: E1128 15:40:44.843493 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f706cb3-8157-46c4-a297-9eb19562105d" containerName="container-00" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.843506 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f706cb3-8157-46c4-a297-9eb19562105d" containerName="container-00" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.843684 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f706cb3-8157-46c4-a297-9eb19562105d" containerName="container-00" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.844437 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.978293 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfrqw\" (UniqueName: \"kubernetes.io/projected/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-kube-api-access-wfrqw\") pod \"crc-debug-6sfkm\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:44 crc kubenswrapper[4817]: I1128 15:40:44.978439 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-host\") pod \"crc-debug-6sfkm\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.079993 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-host\") pod \"crc-debug-6sfkm\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.080122 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-host\") pod \"crc-debug-6sfkm\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.080207 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfrqw\" (UniqueName: \"kubernetes.io/projected/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-kube-api-access-wfrqw\") pod \"crc-debug-6sfkm\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.096704 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfrqw\" (UniqueName: \"kubernetes.io/projected/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-kube-api-access-wfrqw\") pod \"crc-debug-6sfkm\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.161792 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.509302 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" event={"ID":"29779a4b-4f5e-408e-87a8-f302b8ad2b0d","Type":"ContainerStarted","Data":"1158d1123c30741c90fc8ada858fd1d84c3c1adac9fa108508190c45350e664c"} Nov 28 15:40:45 crc kubenswrapper[4817]: I1128 15:40:45.748564 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f706cb3-8157-46c4-a297-9eb19562105d" path="/var/lib/kubelet/pods/1f706cb3-8157-46c4-a297-9eb19562105d/volumes" Nov 28 15:40:46 crc kubenswrapper[4817]: I1128 15:40:46.522226 4817 generic.go:334] "Generic (PLEG): container finished" podID="29779a4b-4f5e-408e-87a8-f302b8ad2b0d" containerID="bd7265df0a05c56c0afa0350ecc5cfe80a5c7498bdeff343b01b3318467f4ab1" exitCode=1 Nov 28 15:40:46 crc kubenswrapper[4817]: I1128 15:40:46.522316 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" event={"ID":"29779a4b-4f5e-408e-87a8-f302b8ad2b0d","Type":"ContainerDied","Data":"bd7265df0a05c56c0afa0350ecc5cfe80a5c7498bdeff343b01b3318467f4ab1"} Nov 28 15:40:46 crc kubenswrapper[4817]: I1128 15:40:46.560296 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9q2hs/crc-debug-6sfkm"] Nov 28 15:40:46 crc kubenswrapper[4817]: I1128 15:40:46.568902 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9q2hs/crc-debug-6sfkm"] Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.682517 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.731523 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfrqw\" (UniqueName: \"kubernetes.io/projected/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-kube-api-access-wfrqw\") pod \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.731978 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-host\") pod \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\" (UID: \"29779a4b-4f5e-408e-87a8-f302b8ad2b0d\") " Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.732091 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-host" (OuterVolumeSpecName: "host") pod "29779a4b-4f5e-408e-87a8-f302b8ad2b0d" (UID: "29779a4b-4f5e-408e-87a8-f302b8ad2b0d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.732711 4817 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-host\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.742963 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-kube-api-access-wfrqw" (OuterVolumeSpecName: "kube-api-access-wfrqw") pod "29779a4b-4f5e-408e-87a8-f302b8ad2b0d" (UID: "29779a4b-4f5e-408e-87a8-f302b8ad2b0d"). InnerVolumeSpecName "kube-api-access-wfrqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.752794 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29779a4b-4f5e-408e-87a8-f302b8ad2b0d" path="/var/lib/kubelet/pods/29779a4b-4f5e-408e-87a8-f302b8ad2b0d/volumes" Nov 28 15:40:47 crc kubenswrapper[4817]: I1128 15:40:47.834905 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfrqw\" (UniqueName: \"kubernetes.io/projected/29779a4b-4f5e-408e-87a8-f302b8ad2b0d-kube-api-access-wfrqw\") on node \"crc\" DevicePath \"\"" Nov 28 15:40:48 crc kubenswrapper[4817]: I1128 15:40:48.562608 4817 scope.go:117] "RemoveContainer" containerID="bd7265df0a05c56c0afa0350ecc5cfe80a5c7498bdeff343b01b3318467f4ab1" Nov 28 15:40:48 crc kubenswrapper[4817]: I1128 15:40:48.562667 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/crc-debug-6sfkm" Nov 28 15:41:52 crc kubenswrapper[4817]: I1128 15:41:52.578334 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/init-config-reloader/0.log" Nov 28 15:41:52 crc kubenswrapper[4817]: I1128 15:41:52.759365 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/init-config-reloader/0.log" Nov 28 15:41:52 crc kubenswrapper[4817]: I1128 15:41:52.793597 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/alertmanager/0.log" Nov 28 15:41:52 crc kubenswrapper[4817]: I1128 15:41:52.838573 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_9f575d2e-8149-4fad-86a3-da8552e0e9e0/config-reloader/0.log" Nov 28 15:41:52 crc kubenswrapper[4817]: I1128 15:41:52.951002 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-api/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.025170 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-evaluator/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.027834 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-listener/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.098977 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e80382bf-f531-4e80-9ff4-2392da846d3e/aodh-notifier/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.236920 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fb9577b84-z8dsl_15698a16-9f4e-4552-b729-27969466063d/barbican-api/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.285297 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fb9577b84-z8dsl_15698a16-9f4e-4552-b729-27969466063d/barbican-api-log/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.469289 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7dd97866bb-qxt2v_7cc501c6-50dd-4f39-b19b-20f10e6b9997/barbican-keystone-listener-log/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.486153 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7dd97866bb-qxt2v_7cc501c6-50dd-4f39-b19b-20f10e6b9997/barbican-keystone-listener/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.588571 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84d84dc879-8hskz_9022c280-17b2-48fc-b818-73797815a819/barbican-worker/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.693553 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84d84dc879-8hskz_9022c280-17b2-48fc-b818-73797815a819/barbican-worker-log/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.791408 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-v6d4q_3200157e-779d-485c-803e-1b0d092c02bb/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.939927 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/ceilometer-central-agent/0.log" Nov 28 15:41:53 crc kubenswrapper[4817]: I1128 15:41:53.951259 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/ceilometer-notification-agent/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.002952 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/proxy-httpd/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.099397 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6f859e2d-de31-407a-86c0-97a3897b1a57/sg-core/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.424664 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8b2cd5c-abbf-4cca-98d0-e0f3db266421/cinder-api/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.457480 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8b2cd5c-abbf-4cca-98d0-e0f3db266421/cinder-api-log/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.646652 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7570a53b-a405-45ee-ab15-ab5f575bb90b/cinder-scheduler/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.693234 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7570a53b-a405-45ee-ab15-ab5f575bb90b/probe/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.791700 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-cjmv5_18ae828a-395c-4300-9460-a063a0322335/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:54 crc kubenswrapper[4817]: I1128 15:41:54.890865 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bcrx8_1a488bc2-89d2-49a6-88a3-0656eb5deb56/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.023834 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-f8k7g_6ecb4997-6b98-4f7b-ad43-e02295ffe723/init/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.197978 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-f8k7g_6ecb4997-6b98-4f7b-ad43-e02295ffe723/init/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.211690 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6mw2q_0452060d-61d9-49c6-b16c-522ded9cb4e8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.258429 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-f8k7g_6ecb4997-6b98-4f7b-ad43-e02295ffe723/dnsmasq-dns/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.436029 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c4f615c7-7cc5-4339-8bdc-e2974582e564/glance-httpd/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.457023 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c4f615c7-7cc5-4339-8bdc-e2974582e564/glance-log/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.628967 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d0451aa8-c32f-46e7-a9d8-574cb7005894/glance-httpd/0.log" Nov 28 15:41:55 crc kubenswrapper[4817]: I1128 15:41:55.637104 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d0451aa8-c32f-46e7-a9d8-574cb7005894/glance-log/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.012585 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6f9fcf49b7-fdlpq_2b12bb76-1c8c-4ea3-afd6-b1ff967ae174/heat-api/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.122061 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7585599b46-rxrsf_bde8e8ec-30e7-4913-83bb-2babaaf11f9a/heat-engine/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.259469 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5696cfb998-r25ks_11d0830f-fa4e-4204-b32d-3ef9553325f1/heat-cfnapi/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.287889 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bkjcd_c817bef2-9ff5-4f69-8a73-1f36128879a6/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.439648 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-zd8vk_62cd0765-6d4e-4e36-96df-8e3cc4bf8a91/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.536472 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c874c964f-4j6hm_b3a6b7c5-6c6d-4ebc-bc5b-b9ab47e95d3d/keystone-api/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.641898 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29405701-7q48x_9ef2895d-d3df-4a6b-abad-c7d1bceb0871/keystone-cron/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.719436 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_95d4474e-82dd-405a-a280-726ce4f894dc/kube-state-metrics/0.log" Nov 28 15:41:56 crc kubenswrapper[4817]: I1128 15:41:56.796293 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-k5cxm_58e1259c-59ec-4bee-867b-b90c25667b33/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:57 crc kubenswrapper[4817]: I1128 15:41:57.051645 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cc96c44df-rqchh_0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9/neutron-api/0.log" Nov 28 15:41:57 crc kubenswrapper[4817]: I1128 15:41:57.177402 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cc96c44df-rqchh_0d59ceee-7dfb-4f7a-8feb-5b165d9de6e9/neutron-httpd/0.log" Nov 28 15:41:57 crc kubenswrapper[4817]: I1128 15:41:57.257930 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jjc6c_f7494198-71fe-4460-a0a5-0737e7fe57dd/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:57 crc kubenswrapper[4817]: I1128 15:41:57.619702 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eb11cfb0-8f52-4361-bbce-b70613a51ca6/nova-api-log/0.log" Nov 28 15:41:57 crc kubenswrapper[4817]: I1128 15:41:57.832739 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_98dff079-658b-4cb4-aa29-42eb2c020705/nova-cell0-conductor-conductor/0.log" Nov 28 15:41:58 crc kubenswrapper[4817]: I1128 15:41:58.065475 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_eb11cfb0-8f52-4361-bbce-b70613a51ca6/nova-api-api/0.log" Nov 28 15:41:58 crc kubenswrapper[4817]: I1128 15:41:58.087934 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_37355c0f-d765-472c-b5d6-beb57bcf9869/nova-cell1-conductor-conductor/0.log" Nov 28 15:41:58 crc kubenswrapper[4817]: I1128 15:41:58.495184 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_514053b0-d172-46fb-ac7c-26bd118d83f0/nova-cell1-novncproxy-novncproxy/0.log" Nov 28 15:41:58 crc kubenswrapper[4817]: I1128 15:41:58.525536 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-g4mz8_8833fd86-9ef4-441d-ada9-bb24bec33153/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:41:58 crc kubenswrapper[4817]: I1128 15:41:58.752058 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_99283929-00ab-45d4-92c9-73cbb67f205f/nova-metadata-log/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.094543 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1fcffa08-235a-4101-bb4f-8933dc9102e3/nova-scheduler-scheduler/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.095735 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d/mysql-bootstrap/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.274058 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d/mysql-bootstrap/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.344273 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ba08a213-ad14-4a0e-a7e5-c09ffa3cde3d/galera/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.501876 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bd723f67-725a-4942-87f2-c646eddcd172/mysql-bootstrap/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.790451 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bd723f67-725a-4942-87f2-c646eddcd172/mysql-bootstrap/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.823977 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bd723f67-725a-4942-87f2-c646eddcd172/galera/0.log" Nov 28 15:41:59 crc kubenswrapper[4817]: I1128 15:41:59.983490 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6df83271-9eeb-4eee-b838-feec85dde399/openstackclient/0.log" Nov 28 15:42:00 crc kubenswrapper[4817]: I1128 15:42:00.424678 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_99283929-00ab-45d4-92c9-73cbb67f205f/nova-metadata-metadata/0.log" Nov 28 15:42:00 crc kubenswrapper[4817]: I1128 15:42:00.788569 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovsdb-server-init/0.log" Nov 28 15:42:00 crc kubenswrapper[4817]: I1128 15:42:00.791366 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9frjn_9ac085c7-55fa-4dd9-9ab6-bfd720716c60/openstack-network-exporter/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.062664 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovsdb-server-init/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.063227 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovs-vswitchd/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.175824 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vz6mt_40df6e7a-e727-4386-bca8-aeff35121dbb/ovsdb-server/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.261844 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pwgz2_0f58e7f4-f831-4f54-91b6-a5e97bf5152c/ovn-controller/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.447651 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-d6ls6_c49884f3-a2a8-49d5-8d6c-c2d7485909ed/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.543965 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_693b7740-a0c9-4430-941a-3f849c254803/openstack-network-exporter/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.552274 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_693b7740-a0c9-4430-941a-3f849c254803/ovn-northd/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.720647 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e3b52366-bcd8-4685-9f36-2adaea4e6cc4/openstack-network-exporter/0.log" Nov 28 15:42:01 crc kubenswrapper[4817]: I1128 15:42:01.804321 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e3b52366-bcd8-4685-9f36-2adaea4e6cc4/ovsdbserver-nb/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.308098 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ff9385f1-3c29-48be-bc80-806380c08646/ovsdbserver-sb/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.350280 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ff9385f1-3c29-48be-bc80-806380c08646/openstack-network-exporter/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.476250 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6746568d76-zwnfd_4164fdcc-c963-4a08-9bd5-2c569d096180/placement-api/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.623562 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/init-config-reloader/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.687558 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6746568d76-zwnfd_4164fdcc-c963-4a08-9bd5-2c569d096180/placement-log/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.854239 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/init-config-reloader/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.872612 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/config-reloader/0.log" Nov 28 15:42:02 crc kubenswrapper[4817]: I1128 15:42:02.949599 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/prometheus/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.023976 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_7afd0bd0-95f4-4a51-b51a-cf1a12b8c80f/thanos-sidecar/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.122351 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a5cbe25b-ed99-4853-a3bd-726f25343f6a/setup-container/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.337313 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a5cbe25b-ed99-4853-a3bd-726f25343f6a/setup-container/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.366397 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a5cbe25b-ed99-4853-a3bd-726f25343f6a/rabbitmq/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.392280 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8a6a65e3-df90-4500-8deb-1dc661954e49/setup-container/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.680944 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-wvj4m_c36cc3fe-b1aa-4858-8c72-1fa223352e09/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.685906 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8a6a65e3-df90-4500-8deb-1dc661954e49/setup-container/0.log" Nov 28 15:42:03 crc kubenswrapper[4817]: I1128 15:42:03.892959 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zzxtk_14f43d11-6db6-461c-9279-22fdd3e286b8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:04 crc kubenswrapper[4817]: I1128 15:42:04.135095 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jxmhd_02fe616c-2dd3-412b-a8e1-9af3150fc92c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:04 crc kubenswrapper[4817]: I1128 15:42:04.228745 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pvgld_0b6cd940-c79b-4bd7-80ff-69bb850709ae/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:04 crc kubenswrapper[4817]: I1128 15:42:04.475196 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-jzgbr_3867da3b-518a-4aa7-80b4-5d6e5f4e7dad/ssh-known-hosts-edpm-deployment/0.log" Nov 28 15:42:04 crc kubenswrapper[4817]: I1128 15:42:04.746333 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85df6b4fc9-7c5nt_72c26e96-3946-44a8-bb20-7786cba96b07/proxy-server/0.log" Nov 28 15:42:04 crc kubenswrapper[4817]: I1128 15:42:04.855679 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85df6b4fc9-7c5nt_72c26e96-3946-44a8-bb20-7786cba96b07/proxy-httpd/0.log" Nov 28 15:42:04 crc kubenswrapper[4817]: I1128 15:42:04.986960 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-fzhb6_f61aa2a2-1ba8-4319-ad02-2c05d6a9f662/swift-ring-rebalance/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.144163 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-auditor/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.230932 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-reaper/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.360152 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-server/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.421975 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/account-replicator/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.461393 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-auditor/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.473882 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8a6a65e3-df90-4500-8deb-1dc661954e49/rabbitmq/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.655141 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-replicator/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.703204 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-updater/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.710781 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-auditor/0.log" Nov 28 15:42:05 crc kubenswrapper[4817]: I1128 15:42:05.855752 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/container-server/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.081904 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-expirer/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.133866 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-server/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.137763 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-replicator/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.227492 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/object-updater/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.277354 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/rsync/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.346394 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_42276c9f-fac8-4794-a69c-02178767e5b1/swift-recon-cron/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.558054 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2cnl8_aa1fe5c5-0d23-47fb-990d-480cd18970b8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:06 crc kubenswrapper[4817]: I1128 15:42:06.600556 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-622cj_f9a2d436-9de0-46c8-8a53-073060648992/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 15:42:15 crc kubenswrapper[4817]: I1128 15:42:15.036043 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_789d2be2-8615-428f-bc54-68d019823a99/memcached/0.log" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.112386 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tf8rc"] Nov 28 15:42:20 crc kubenswrapper[4817]: E1128 15:42:20.113548 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29779a4b-4f5e-408e-87a8-f302b8ad2b0d" containerName="container-00" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.113562 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="29779a4b-4f5e-408e-87a8-f302b8ad2b0d" containerName="container-00" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.113956 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="29779a4b-4f5e-408e-87a8-f302b8ad2b0d" containerName="container-00" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.116156 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.132582 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tf8rc"] Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.178519 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wpmb\" (UniqueName: \"kubernetes.io/projected/8d4bd1e7-bb30-445b-a04f-1a89322f8259-kube-api-access-9wpmb\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.178661 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-utilities\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.178739 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-catalog-content\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.280363 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wpmb\" (UniqueName: \"kubernetes.io/projected/8d4bd1e7-bb30-445b-a04f-1a89322f8259-kube-api-access-9wpmb\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.280475 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-utilities\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.280519 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-catalog-content\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.280972 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-catalog-content\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.281050 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-utilities\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.303436 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wpmb\" (UniqueName: \"kubernetes.io/projected/8d4bd1e7-bb30-445b-a04f-1a89322f8259-kube-api-access-9wpmb\") pod \"community-operators-tf8rc\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.446111 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:20 crc kubenswrapper[4817]: I1128 15:42:20.964425 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tf8rc"] Nov 28 15:42:20 crc kubenswrapper[4817]: W1128 15:42:20.976615 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d4bd1e7_bb30_445b_a04f_1a89322f8259.slice/crio-61deffb6e9305d095aec9c0916a9d9055dd08bf70cd9c0d0d3bfff3d45c18f02 WatchSource:0}: Error finding container 61deffb6e9305d095aec9c0916a9d9055dd08bf70cd9c0d0d3bfff3d45c18f02: Status 404 returned error can't find the container with id 61deffb6e9305d095aec9c0916a9d9055dd08bf70cd9c0d0d3bfff3d45c18f02 Nov 28 15:42:21 crc kubenswrapper[4817]: I1128 15:42:21.556857 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerStarted","Data":"61deffb6e9305d095aec9c0916a9d9055dd08bf70cd9c0d0d3bfff3d45c18f02"} Nov 28 15:42:22 crc kubenswrapper[4817]: I1128 15:42:22.568693 4817 generic.go:334] "Generic (PLEG): container finished" podID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerID="c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8" exitCode=0 Nov 28 15:42:22 crc kubenswrapper[4817]: I1128 15:42:22.569154 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerDied","Data":"c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8"} Nov 28 15:42:23 crc kubenswrapper[4817]: I1128 15:42:23.580525 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerStarted","Data":"9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1"} Nov 28 15:42:24 crc kubenswrapper[4817]: I1128 15:42:24.045214 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:42:24 crc kubenswrapper[4817]: I1128 15:42:24.045594 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:42:24 crc kubenswrapper[4817]: I1128 15:42:24.591110 4817 generic.go:334] "Generic (PLEG): container finished" podID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerID="9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1" exitCode=0 Nov 28 15:42:24 crc kubenswrapper[4817]: I1128 15:42:24.591156 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerDied","Data":"9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1"} Nov 28 15:42:25 crc kubenswrapper[4817]: I1128 15:42:25.602392 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerStarted","Data":"3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc"} Nov 28 15:42:25 crc kubenswrapper[4817]: I1128 15:42:25.625868 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tf8rc" podStartSLOduration=3.196348681 podStartE2EDuration="5.625842085s" podCreationTimestamp="2025-11-28 15:42:20 +0000 UTC" firstStartedPulling="2025-11-28 15:42:22.572444761 +0000 UTC m=+4365.160423027" lastFinishedPulling="2025-11-28 15:42:25.001938165 +0000 UTC m=+4367.589916431" observedRunningTime="2025-11-28 15:42:25.619326294 +0000 UTC m=+4368.207304560" watchObservedRunningTime="2025-11-28 15:42:25.625842085 +0000 UTC m=+4368.213820351" Nov 28 15:42:30 crc kubenswrapper[4817]: I1128 15:42:30.446430 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:30 crc kubenswrapper[4817]: I1128 15:42:30.447080 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:31 crc kubenswrapper[4817]: I1128 15:42:31.400833 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:31 crc kubenswrapper[4817]: I1128 15:42:31.456509 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:31 crc kubenswrapper[4817]: I1128 15:42:31.652082 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tf8rc"] Nov 28 15:42:32 crc kubenswrapper[4817]: I1128 15:42:32.672052 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tf8rc" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="registry-server" containerID="cri-o://3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc" gracePeriod=2 Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.469999 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.662456 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-utilities\") pod \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.662585 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wpmb\" (UniqueName: \"kubernetes.io/projected/8d4bd1e7-bb30-445b-a04f-1a89322f8259-kube-api-access-9wpmb\") pod \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.662676 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-catalog-content\") pod \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\" (UID: \"8d4bd1e7-bb30-445b-a04f-1a89322f8259\") " Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.664343 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-utilities" (OuterVolumeSpecName: "utilities") pod "8d4bd1e7-bb30-445b-a04f-1a89322f8259" (UID: "8d4bd1e7-bb30-445b-a04f-1a89322f8259"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.676073 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d4bd1e7-bb30-445b-a04f-1a89322f8259-kube-api-access-9wpmb" (OuterVolumeSpecName: "kube-api-access-9wpmb") pod "8d4bd1e7-bb30-445b-a04f-1a89322f8259" (UID: "8d4bd1e7-bb30-445b-a04f-1a89322f8259"). InnerVolumeSpecName "kube-api-access-9wpmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.693406 4817 generic.go:334] "Generic (PLEG): container finished" podID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerID="3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc" exitCode=0 Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.693751 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf8rc" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.693795 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerDied","Data":"3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc"} Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.694231 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf8rc" event={"ID":"8d4bd1e7-bb30-445b-a04f-1a89322f8259","Type":"ContainerDied","Data":"61deffb6e9305d095aec9c0916a9d9055dd08bf70cd9c0d0d3bfff3d45c18f02"} Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.694466 4817 scope.go:117] "RemoveContainer" containerID="3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.722859 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d4bd1e7-bb30-445b-a04f-1a89322f8259" (UID: "8d4bd1e7-bb30-445b-a04f-1a89322f8259"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.733213 4817 scope.go:117] "RemoveContainer" containerID="9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.765998 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wpmb\" (UniqueName: \"kubernetes.io/projected/8d4bd1e7-bb30-445b-a04f-1a89322f8259-kube-api-access-9wpmb\") on node \"crc\" DevicePath \"\"" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.766035 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.766047 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d4bd1e7-bb30-445b-a04f-1a89322f8259-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.775661 4817 scope.go:117] "RemoveContainer" containerID="c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.807430 4817 scope.go:117] "RemoveContainer" containerID="3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc" Nov 28 15:42:33 crc kubenswrapper[4817]: E1128 15:42:33.807838 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc\": container with ID starting with 3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc not found: ID does not exist" containerID="3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.807881 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc"} err="failed to get container status \"3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc\": rpc error: code = NotFound desc = could not find container \"3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc\": container with ID starting with 3658886228b2a2da0a1f3a932a6cbc70baf439089681d2cadde0f1e12610d0fc not found: ID does not exist" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.807910 4817 scope.go:117] "RemoveContainer" containerID="9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1" Nov 28 15:42:33 crc kubenswrapper[4817]: E1128 15:42:33.808239 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1\": container with ID starting with 9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1 not found: ID does not exist" containerID="9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.808270 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1"} err="failed to get container status \"9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1\": rpc error: code = NotFound desc = could not find container \"9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1\": container with ID starting with 9ff31d840a7ace235b93556e30828efdbd33274ddbc05b8454bfe0904dc7e2b1 not found: ID does not exist" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.808288 4817 scope.go:117] "RemoveContainer" containerID="c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8" Nov 28 15:42:33 crc kubenswrapper[4817]: E1128 15:42:33.808638 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8\": container with ID starting with c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8 not found: ID does not exist" containerID="c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8" Nov 28 15:42:33 crc kubenswrapper[4817]: I1128 15:42:33.808663 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8"} err="failed to get container status \"c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8\": rpc error: code = NotFound desc = could not find container \"c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8\": container with ID starting with c14814e4ae9af2e3fe130d60d076264fdf6d7272037418be9f1be058967368a8 not found: ID does not exist" Nov 28 15:42:34 crc kubenswrapper[4817]: I1128 15:42:34.021456 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tf8rc"] Nov 28 15:42:34 crc kubenswrapper[4817]: I1128 15:42:34.033166 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tf8rc"] Nov 28 15:42:35 crc kubenswrapper[4817]: I1128 15:42:35.749167 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" path="/var/lib/kubelet/pods/8d4bd1e7-bb30-445b-a04f-1a89322f8259/volumes" Nov 28 15:42:36 crc kubenswrapper[4817]: I1128 15:42:36.993005 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-8lb8j_59388928-9001-4abd-9361-d4045b4d027b/kube-rbac-proxy/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.012423 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-8lb8j_59388928-9001-4abd-9361-d4045b4d027b/manager/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.174071 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-75m4b_b0544c31-6ad6-4a76-84da-c180f89ba5ee/kube-rbac-proxy/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.233801 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-75m4b_b0544c31-6ad6-4a76-84da-c180f89ba5ee/manager/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.352564 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/util/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.536816 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/pull/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.583336 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/util/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.598909 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/pull/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.708226 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/util/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.785659 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/extract/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.839627 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dd62600bcdc119f05aaa90124f3e4d7bf0f3fb184c3a8c276e407c1d55r5vcc_ef879694-f24b-49a3-98d2-82a71aa98d11/pull/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.954036 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-8cjxb_025353e3-a121-42b5-b283-727b6eb24b7c/kube-rbac-proxy/0.log" Nov 28 15:42:37 crc kubenswrapper[4817]: I1128 15:42:37.969309 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-8cjxb_025353e3-a121-42b5-b283-727b6eb24b7c/manager/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.095907 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-4fj4z_29cc111e-478d-48c6-a89c-4050114b5d37/kube-rbac-proxy/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.217425 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-4fj4z_29cc111e-478d-48c6-a89c-4050114b5d37/manager/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.291226 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-k7tbk_02f94412-5d1b-401d-a9fa-211df97a2710/kube-rbac-proxy/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.412358 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-k7tbk_02f94412-5d1b-401d-a9fa-211df97a2710/manager/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.473427 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-j2vsj_e04c0207-7d18-437a-81a3-51ca263e492b/manager/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.501322 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-j2vsj_e04c0207-7d18-437a-81a3-51ca263e492b/kube-rbac-proxy/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.669077 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-w9qqs_a55aea0f-52dc-4403-8067-2372b24cdf83/kube-rbac-proxy/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.845526 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-vh4rl_4598ac96-dead-47cc-b063-22d3197689a4/kube-rbac-proxy/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.883124 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-w9qqs_a55aea0f-52dc-4403-8067-2372b24cdf83/manager/0.log" Nov 28 15:42:38 crc kubenswrapper[4817]: I1128 15:42:38.912010 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-vh4rl_4598ac96-dead-47cc-b063-22d3197689a4/manager/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.074113 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4phnt_694e1cb6-aa3a-4909-b220-99d42dbf1e49/kube-rbac-proxy/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.171617 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4phnt_694e1cb6-aa3a-4909-b220-99d42dbf1e49/manager/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.274899 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-tbssr_be902d34-5e5e-48be-9a6b-d9e53661f367/kube-rbac-proxy/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.332267 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-tbssr_be902d34-5e5e-48be-9a6b-d9e53661f367/manager/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.392087 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-87bqn_19220366-8f0a-4108-9093-964f54a4642f/kube-rbac-proxy/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.450832 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t48qk"] Nov 28 15:42:39 crc kubenswrapper[4817]: E1128 15:42:39.451245 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="registry-server" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.451262 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="registry-server" Nov 28 15:42:39 crc kubenswrapper[4817]: E1128 15:42:39.451284 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="extract-utilities" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.451293 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="extract-utilities" Nov 28 15:42:39 crc kubenswrapper[4817]: E1128 15:42:39.451321 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="extract-content" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.451331 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="extract-content" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.451539 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d4bd1e7-bb30-445b-a04f-1a89322f8259" containerName="registry-server" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.453040 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.463459 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t48qk"] Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.535472 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-87bqn_19220366-8f0a-4108-9093-964f54a4642f/manager/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.571168 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-catalog-content\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.571232 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjsxt\" (UniqueName: \"kubernetes.io/projected/6745e75b-636e-4176-92c8-23d37e960866-kube-api-access-gjsxt\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.571291 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-utilities\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.673862 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-catalog-content\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.673913 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjsxt\" (UniqueName: \"kubernetes.io/projected/6745e75b-636e-4176-92c8-23d37e960866-kube-api-access-gjsxt\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.673953 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-utilities\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.674529 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-catalog-content\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.674574 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-utilities\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.679804 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-pmw8p_2f81ca1e-831f-4a15-9ea9-a81364b534fd/kube-rbac-proxy/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.703609 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjsxt\" (UniqueName: \"kubernetes.io/projected/6745e75b-636e-4176-92c8-23d37e960866-kube-api-access-gjsxt\") pod \"redhat-operators-t48qk\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.724638 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-pmw8p_2f81ca1e-831f-4a15-9ea9-a81364b534fd/manager/0.log" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.776051 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:39 crc kubenswrapper[4817]: I1128 15:42:39.852756 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-wzw2n_abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29/kube-rbac-proxy/0.log" Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.151604 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-wzw2n_abcdec0b-5fe1-4b64-a090-a0f5f3ae3f29/manager/0.log" Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.245340 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-gv25c_18412c56-76ce-48a3-81f2-3cbc4d090df5/kube-rbac-proxy/0.log" Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.305141 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-gv25c_18412c56-76ce-48a3-81f2-3cbc4d090df5/manager/0.log" Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.329089 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t48qk"] Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.411708 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm_35934979-2374-4a86-b9db-7effb7624f40/kube-rbac-proxy/0.log" Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.566478 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bn87fm_35934979-2374-4a86-b9db-7effb7624f40/manager/0.log" Nov 28 15:42:40 crc kubenswrapper[4817]: I1128 15:42:40.771902 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerStarted","Data":"6f1304338e258157d38958af5e3325aad0df701e932c284ac9378dfaa57e971c"} Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.082672 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56ffd7cc6d-rw9vh_cb0ec512-1d2d-4487-8d7e-c412ea2b4da6/operator/0.log" Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.270682 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mt7jk_3efb685f-68cf-4455-9e24-ac2d54d45dab/registry-server/0.log" Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.744004 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8fqs4_5343b3a1-d2c2-4b26-806b-a7661275d20f/kube-rbac-proxy/0.log" Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.783243 4817 generic.go:334] "Generic (PLEG): container finished" podID="6745e75b-636e-4176-92c8-23d37e960866" containerID="32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b" exitCode=0 Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.783291 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerDied","Data":"32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b"} Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.876324 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-s52pg_6ec7b27a-79e7-402c-8b73-091c9c20a1db/manager/0.log" Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.889911 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-s52pg_6ec7b27a-79e7-402c-8b73-091c9c20a1db/kube-rbac-proxy/0.log" Nov 28 15:42:41 crc kubenswrapper[4817]: I1128 15:42:41.983345 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-8fqs4_5343b3a1-d2c2-4b26-806b-a7661275d20f/manager/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.170596 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jr75r_d9e5f208-4552-48ff-af77-901fa7c76e42/operator/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.176288 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64d6f76c7b-5smh2_c6203ca7-1ea0-44ae-8c4f-cc3637ae0af5/manager/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.236649 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-n96rm_a023b80a-0d1c-47c6-b904-014d659e559a/kube-rbac-proxy/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.242653 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-n96rm_a023b80a-0d1c-47c6-b904-014d659e559a/manager/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.351644 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/kube-rbac-proxy/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.470623 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-xpvwd_e8b4fc65-1b95-41fb-a95e-6e102c452691/kube-rbac-proxy/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.494790 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7b8977cb-jkx7n_cafcb9ae-aa1a-490e-b27d-656bf4b8508d/manager/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.631130 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-xpvwd_e8b4fc65-1b95-41fb-a95e-6e102c452691/manager/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.686069 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-26v9h_77766fad-8504-4053-a3f6-78591e568ce5/kube-rbac-proxy/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.686135 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-26v9h_77766fad-8504-4053-a3f6-78591e568ce5/manager/0.log" Nov 28 15:42:42 crc kubenswrapper[4817]: I1128 15:42:42.793905 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerStarted","Data":"cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76"} Nov 28 15:42:48 crc kubenswrapper[4817]: I1128 15:42:48.848331 4817 generic.go:334] "Generic (PLEG): container finished" podID="6745e75b-636e-4176-92c8-23d37e960866" containerID="cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76" exitCode=0 Nov 28 15:42:48 crc kubenswrapper[4817]: I1128 15:42:48.848811 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerDied","Data":"cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76"} Nov 28 15:42:51 crc kubenswrapper[4817]: I1128 15:42:51.876408 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerStarted","Data":"65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a"} Nov 28 15:42:51 crc kubenswrapper[4817]: I1128 15:42:51.901576 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t48qk" podStartSLOduration=3.926795643 podStartE2EDuration="12.901556733s" podCreationTimestamp="2025-11-28 15:42:39 +0000 UTC" firstStartedPulling="2025-11-28 15:42:41.791579122 +0000 UTC m=+4384.379557388" lastFinishedPulling="2025-11-28 15:42:50.766340212 +0000 UTC m=+4393.354318478" observedRunningTime="2025-11-28 15:42:51.897089282 +0000 UTC m=+4394.485067548" watchObservedRunningTime="2025-11-28 15:42:51.901556733 +0000 UTC m=+4394.489534999" Nov 28 15:42:54 crc kubenswrapper[4817]: I1128 15:42:54.045793 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:42:54 crc kubenswrapper[4817]: I1128 15:42:54.046352 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:42:59 crc kubenswrapper[4817]: I1128 15:42:59.776236 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:59 crc kubenswrapper[4817]: I1128 15:42:59.776740 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:42:59 crc kubenswrapper[4817]: I1128 15:42:59.822540 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:43:00 crc kubenswrapper[4817]: I1128 15:43:00.006101 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:43:00 crc kubenswrapper[4817]: I1128 15:43:00.062413 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t48qk"] Nov 28 15:43:01 crc kubenswrapper[4817]: I1128 15:43:01.972324 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t48qk" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="registry-server" containerID="cri-o://65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a" gracePeriod=2 Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.451618 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-kvssf_55dbc750-3c63-44f9-ada1-aa282adb3562/control-plane-machine-set-operator/0.log" Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.957087 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.983545 4817 generic.go:334] "Generic (PLEG): container finished" podID="6745e75b-636e-4176-92c8-23d37e960866" containerID="65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a" exitCode=0 Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.983599 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerDied","Data":"65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a"} Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.983637 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48qk" event={"ID":"6745e75b-636e-4176-92c8-23d37e960866","Type":"ContainerDied","Data":"6f1304338e258157d38958af5e3325aad0df701e932c284ac9378dfaa57e971c"} Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.983663 4817 scope.go:117] "RemoveContainer" containerID="65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a" Nov 28 15:43:02 crc kubenswrapper[4817]: I1128 15:43:02.983872 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48qk" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.011584 4817 scope.go:117] "RemoveContainer" containerID="cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.031807 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-skq7g_e20cea1b-1075-4334-a1b6-3634ee774393/machine-api-operator/0.log" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.034624 4817 scope.go:117] "RemoveContainer" containerID="32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.088111 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-skq7g_e20cea1b-1075-4334-a1b6-3634ee774393/kube-rbac-proxy/0.log" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.133841 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-catalog-content\") pod \"6745e75b-636e-4176-92c8-23d37e960866\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.134260 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-utilities\") pod \"6745e75b-636e-4176-92c8-23d37e960866\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.134420 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjsxt\" (UniqueName: \"kubernetes.io/projected/6745e75b-636e-4176-92c8-23d37e960866-kube-api-access-gjsxt\") pod \"6745e75b-636e-4176-92c8-23d37e960866\" (UID: \"6745e75b-636e-4176-92c8-23d37e960866\") " Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.135121 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-utilities" (OuterVolumeSpecName: "utilities") pod "6745e75b-636e-4176-92c8-23d37e960866" (UID: "6745e75b-636e-4176-92c8-23d37e960866"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.140691 4817 scope.go:117] "RemoveContainer" containerID="65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.140778 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6745e75b-636e-4176-92c8-23d37e960866-kube-api-access-gjsxt" (OuterVolumeSpecName: "kube-api-access-gjsxt") pod "6745e75b-636e-4176-92c8-23d37e960866" (UID: "6745e75b-636e-4176-92c8-23d37e960866"). InnerVolumeSpecName "kube-api-access-gjsxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:43:03 crc kubenswrapper[4817]: E1128 15:43:03.145920 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a\": container with ID starting with 65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a not found: ID does not exist" containerID="65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.145970 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a"} err="failed to get container status \"65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a\": rpc error: code = NotFound desc = could not find container \"65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a\": container with ID starting with 65629d3b3119b6c42d0180837bbdf0a8beacc691de197ff1f551a4b97969b53a not found: ID does not exist" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.146001 4817 scope.go:117] "RemoveContainer" containerID="cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76" Nov 28 15:43:03 crc kubenswrapper[4817]: E1128 15:43:03.146464 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76\": container with ID starting with cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76 not found: ID does not exist" containerID="cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.146575 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76"} err="failed to get container status \"cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76\": rpc error: code = NotFound desc = could not find container \"cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76\": container with ID starting with cc95337f636f1728a13a5a991e098990091302ac5c4bde9813b2849cde49ca76 not found: ID does not exist" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.146666 4817 scope.go:117] "RemoveContainer" containerID="32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b" Nov 28 15:43:03 crc kubenswrapper[4817]: E1128 15:43:03.147182 4817 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b\": container with ID starting with 32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b not found: ID does not exist" containerID="32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.147238 4817 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b"} err="failed to get container status \"32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b\": rpc error: code = NotFound desc = could not find container \"32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b\": container with ID starting with 32bddf6bcc680200a0698f2995a7ff037e69736eb4a6de4069304eb1288d4c6b not found: ID does not exist" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.236753 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.237050 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjsxt\" (UniqueName: \"kubernetes.io/projected/6745e75b-636e-4176-92c8-23d37e960866-kube-api-access-gjsxt\") on node \"crc\" DevicePath \"\"" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.249616 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6745e75b-636e-4176-92c8-23d37e960866" (UID: "6745e75b-636e-4176-92c8-23d37e960866"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.330900 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t48qk"] Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.338627 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6745e75b-636e-4176-92c8-23d37e960866-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.345099 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t48qk"] Nov 28 15:43:03 crc kubenswrapper[4817]: I1128 15:43:03.749949 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6745e75b-636e-4176-92c8-23d37e960866" path="/var/lib/kubelet/pods/6745e75b-636e-4176-92c8-23d37e960866/volumes" Nov 28 15:43:15 crc kubenswrapper[4817]: I1128 15:43:15.363517 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-d2rzq_d33ac74c-861f-4b72-8ba8-38e14244b510/cert-manager-controller/0.log" Nov 28 15:43:15 crc kubenswrapper[4817]: I1128 15:43:15.548973 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-fmdv9_d6857358-29dc-4bfa-89cc-503c7d452236/cert-manager-cainjector/0.log" Nov 28 15:43:15 crc kubenswrapper[4817]: I1128 15:43:15.584055 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xfd76_aee642f7-2003-406e-b5fd-b5a9b6c6e1fe/cert-manager-webhook/0.log" Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.045056 4817 patch_prober.go:28] interesting pod/machine-config-daemon-6kh49 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.045761 4817 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.045814 4817 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.046604 4817 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41"} pod="openshift-machine-config-operator/machine-config-daemon-6kh49" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.046673 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerName="machine-config-daemon" containerID="cri-o://683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" gracePeriod=600 Nov 28 15:43:24 crc kubenswrapper[4817]: E1128 15:43:24.194821 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.201816 4817 generic.go:334] "Generic (PLEG): container finished" podID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" exitCode=0 Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.201861 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerDied","Data":"683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41"} Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.201905 4817 scope.go:117] "RemoveContainer" containerID="898fa3fe56fad6eb55fca0346a27b58f52616da04be43eac8a66a01390e66f20" Nov 28 15:43:24 crc kubenswrapper[4817]: I1128 15:43:24.202550 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:43:24 crc kubenswrapper[4817]: E1128 15:43:24.202950 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:43:24 crc kubenswrapper[4817]: E1128 15:43:24.206312 4817 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f2c731e_6fa9_4e0c_81dd_115f13fa4dfc.slice/crio-conmon-683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f2c731e_6fa9_4e0c_81dd_115f13fa4dfc.slice/crio-683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41.scope\": RecentStats: unable to find data in memory cache]" Nov 28 15:43:29 crc kubenswrapper[4817]: I1128 15:43:29.455985 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-hzvx2_f9dc9469-cf27-4d53-ae6d-a4df72d8af78/nmstate-console-plugin/0.log" Nov 28 15:43:29 crc kubenswrapper[4817]: I1128 15:43:29.642248 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kgbh6_feed3d1f-a072-4ccc-b47a-3a5b48b44177/nmstate-handler/0.log" Nov 28 15:43:29 crc kubenswrapper[4817]: I1128 15:43:29.698882 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-wgrqq_90fcff6a-9f08-479c-8e5b-2745700bc8f1/nmstate-metrics/0.log" Nov 28 15:43:29 crc kubenswrapper[4817]: I1128 15:43:29.702146 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-wgrqq_90fcff6a-9f08-479c-8e5b-2745700bc8f1/kube-rbac-proxy/0.log" Nov 28 15:43:29 crc kubenswrapper[4817]: I1128 15:43:29.930468 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-b4g4f_68d8ec1c-4246-402e-8da4-53dfe24edbca/nmstate-operator/0.log" Nov 28 15:43:29 crc kubenswrapper[4817]: I1128 15:43:29.956595 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-pm78n_8d2164f8-732d-4650-b92a-3fb677b51c25/nmstate-webhook/0.log" Nov 28 15:43:38 crc kubenswrapper[4817]: I1128 15:43:38.737948 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:43:38 crc kubenswrapper[4817]: E1128 15:43:38.740344 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:43:45 crc kubenswrapper[4817]: I1128 15:43:45.664044 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-4wsbn_7f4b2cee-aa3d-43ca-adbd-42dee238c7bc/kube-rbac-proxy/0.log" Nov 28 15:43:45 crc kubenswrapper[4817]: I1128 15:43:45.752957 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-4wsbn_7f4b2cee-aa3d-43ca-adbd-42dee238c7bc/controller/0.log" Nov 28 15:43:45 crc kubenswrapper[4817]: I1128 15:43:45.847408 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.003463 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.033029 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.056393 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.100030 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.258085 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.276778 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.334622 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.372310 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.561540 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-reloader/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.570989 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/controller/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.590799 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-frr-files/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.599567 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/cp-metrics/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.759425 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/kube-rbac-proxy/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.790519 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/frr-metrics/0.log" Nov 28 15:43:46 crc kubenswrapper[4817]: I1128 15:43:46.855489 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/kube-rbac-proxy-frr/0.log" Nov 28 15:43:47 crc kubenswrapper[4817]: I1128 15:43:47.011950 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/reloader/0.log" Nov 28 15:43:47 crc kubenswrapper[4817]: I1128 15:43:47.064826 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-4r56x_8f400971-fd8a-4c48-82bd-33f47438bcb0/frr-k8s-webhook-server/0.log" Nov 28 15:43:47 crc kubenswrapper[4817]: I1128 15:43:47.294888 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d674c8684-kk2kj_aebf367a-7ec1-4675-b01f-3d54d2ba0613/manager/0.log" Nov 28 15:43:47 crc kubenswrapper[4817]: I1128 15:43:47.477946 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74f997fcc-g6nm4_5980acb2-a5d3-44fa-8bf0-23e36cde74a4/webhook-server/0.log" Nov 28 15:43:47 crc kubenswrapper[4817]: I1128 15:43:47.583150 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkbsn_f737005e-fcc0-4f24-bdbb-99d7f22afef6/kube-rbac-proxy/0.log" Nov 28 15:43:48 crc kubenswrapper[4817]: I1128 15:43:48.177737 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkbsn_f737005e-fcc0-4f24-bdbb-99d7f22afef6/speaker/0.log" Nov 28 15:43:48 crc kubenswrapper[4817]: I1128 15:43:48.410310 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfq4b_39068359-e185-4366-9378-718043d84ba2/frr/0.log" Nov 28 15:43:52 crc kubenswrapper[4817]: I1128 15:43:52.736854 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:43:52 crc kubenswrapper[4817]: E1128 15:43:52.737738 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:44:03 crc kubenswrapper[4817]: I1128 15:44:03.603563 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/util/0.log" Nov 28 15:44:03 crc kubenswrapper[4817]: I1128 15:44:03.768144 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/util/0.log" Nov 28 15:44:03 crc kubenswrapper[4817]: I1128 15:44:03.777495 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/pull/0.log" Nov 28 15:44:03 crc kubenswrapper[4817]: I1128 15:44:03.827623 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/pull/0.log" Nov 28 15:44:04 crc kubenswrapper[4817]: I1128 15:44:04.603892 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/extract/0.log" Nov 28 15:44:04 crc kubenswrapper[4817]: I1128 15:44:04.637143 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/pull/0.log" Nov 28 15:44:04 crc kubenswrapper[4817]: I1128 15:44:04.656737 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frxrdt_83c91d8d-d9d8-4952-ba8b-b57f1d003806/util/0.log" Nov 28 15:44:04 crc kubenswrapper[4817]: I1128 15:44:04.814390 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/util/0.log" Nov 28 15:44:04 crc kubenswrapper[4817]: I1128 15:44:04.998810 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/pull/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.005586 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/util/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.017492 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/pull/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.157074 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/util/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.270701 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/extract/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.517639 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210v5bsz_dc724ad5-cdc7-401d-b139-7aa1ff965410/pull/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.633611 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/util/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.787666 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/util/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.832130 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/pull/0.log" Nov 28 15:44:05 crc kubenswrapper[4817]: I1128 15:44:05.876337 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/pull/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.001741 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/util/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.015415 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/pull/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.020975 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8377hn5_645c8f43-6461-4c60-8fca-ccdde864e321/extract/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.191299 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-utilities/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.364913 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-utilities/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.379267 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-content/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.385126 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-content/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.617122 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-utilities/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.631868 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/extract-content/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.737613 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:44:06 crc kubenswrapper[4817]: E1128 15:44:06.737854 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.797043 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-utilities/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.985075 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-content/0.log" Nov 28 15:44:06 crc kubenswrapper[4817]: I1128 15:44:06.997666 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-utilities/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.083749 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-content/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.249022 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r22b9_7e0d724a-6b6d-4fee-8481-7649c214af98/registry-server/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.291649 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-content/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.295776 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/extract-utilities/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.491971 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hxjk5_9482d966-5b89-4687-9b4a-29eefc683dcd/marketplace-operator/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.533170 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-utilities/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.808174 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-utilities/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.839456 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-content/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.846097 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-content/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.869907 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d7r7v_f69b168e-d53d-476a-b6bf-2dbf64d79cca/registry-server/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.963351 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-utilities/0.log" Nov 28 15:44:07 crc kubenswrapper[4817]: I1128 15:44:07.991152 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/extract-content/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.057158 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-utilities/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.154978 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-q2tpd_0ef612d5-ca83-442b-bf1c-6f4e16d10be2/registry-server/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.263393 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-utilities/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.268128 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-content/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.273865 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-content/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.445297 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-utilities/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.455250 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/extract-content/0.log" Nov 28 15:44:08 crc kubenswrapper[4817]: I1128 15:44:08.928042 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h8h9j_0c4cbeb0-f40f-4b47-9e11-21878c2f0848/registry-server/0.log" Nov 28 15:44:17 crc kubenswrapper[4817]: I1128 15:44:17.746681 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:44:17 crc kubenswrapper[4817]: E1128 15:44:17.747512 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:44:21 crc kubenswrapper[4817]: I1128 15:44:21.755769 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-jvrtq_ad4b53c3-5d16-412b-ba96-f4ba1a54624a/prometheus-operator/0.log" Nov 28 15:44:21 crc kubenswrapper[4817]: I1128 15:44:21.945109 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c6f9b487b-gf8tw_4f9f00a0-1c1f-436e-b59b-7f28c0ef7eaf/prometheus-operator-admission-webhook/0.log" Nov 28 15:44:21 crc kubenswrapper[4817]: I1128 15:44:21.979541 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c6f9b487b-qbqvm_597ad435-7283-469d-8c9a-b0e9a032a433/prometheus-operator-admission-webhook/0.log" Nov 28 15:44:22 crc kubenswrapper[4817]: I1128 15:44:22.108853 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-7mc44_9bd8cd77-0f23-40cc-87aa-19ef553d9565/operator/0.log" Nov 28 15:44:22 crc kubenswrapper[4817]: I1128 15:44:22.214564 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-gcb2c_334ea1c5-47bb-4c1d-b2ae-939ec2b4f76f/perses-operator/0.log" Nov 28 15:44:30 crc kubenswrapper[4817]: I1128 15:44:30.737661 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:44:30 crc kubenswrapper[4817]: E1128 15:44:30.738453 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.637904 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-52hrt"] Nov 28 15:44:44 crc kubenswrapper[4817]: E1128 15:44:44.638983 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="extract-utilities" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.639001 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="extract-utilities" Nov 28 15:44:44 crc kubenswrapper[4817]: E1128 15:44:44.639018 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="extract-content" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.639026 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="extract-content" Nov 28 15:44:44 crc kubenswrapper[4817]: E1128 15:44:44.639067 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="registry-server" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.639075 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="registry-server" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.639325 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="6745e75b-636e-4176-92c8-23d37e960866" containerName="registry-server" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.641108 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.669449 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52hrt"] Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.778755 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x87ln\" (UniqueName: \"kubernetes.io/projected/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-kube-api-access-x87ln\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.778847 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-catalog-content\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.778944 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-utilities\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.881179 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-utilities\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.881469 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x87ln\" (UniqueName: \"kubernetes.io/projected/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-kube-api-access-x87ln\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.881611 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-catalog-content\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.882187 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-utilities\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.882947 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-catalog-content\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.903769 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x87ln\" (UniqueName: \"kubernetes.io/projected/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-kube-api-access-x87ln\") pod \"redhat-marketplace-52hrt\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:44 crc kubenswrapper[4817]: I1128 15:44:44.966331 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:45 crc kubenswrapper[4817]: I1128 15:44:45.517343 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-52hrt"] Nov 28 15:44:45 crc kubenswrapper[4817]: I1128 15:44:45.737194 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:44:45 crc kubenswrapper[4817]: E1128 15:44:45.738522 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:44:45 crc kubenswrapper[4817]: I1128 15:44:45.957153 4817 generic.go:334] "Generic (PLEG): container finished" podID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerID="77e1849b262848d3716506efc6a44ef7ef34cae366e8e9fb889fb17904fc1c5d" exitCode=0 Nov 28 15:44:45 crc kubenswrapper[4817]: I1128 15:44:45.957292 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52hrt" event={"ID":"ab76693e-7ecc-474f-8d48-8c77eb2bdd75","Type":"ContainerDied","Data":"77e1849b262848d3716506efc6a44ef7ef34cae366e8e9fb889fb17904fc1c5d"} Nov 28 15:44:45 crc kubenswrapper[4817]: I1128 15:44:45.957457 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52hrt" event={"ID":"ab76693e-7ecc-474f-8d48-8c77eb2bdd75","Type":"ContainerStarted","Data":"609dbdcbde830e23f1221efdd029afe5d627f7067c927404678fd14c0761bc64"} Nov 28 15:44:45 crc kubenswrapper[4817]: I1128 15:44:45.959833 4817 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 15:44:47 crc kubenswrapper[4817]: I1128 15:44:47.981924 4817 generic.go:334] "Generic (PLEG): container finished" podID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerID="a1b2f0987e9162c2b07f706325223d69cef9715ecc8c25e793e8743051442e88" exitCode=0 Nov 28 15:44:47 crc kubenswrapper[4817]: I1128 15:44:47.982861 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52hrt" event={"ID":"ab76693e-7ecc-474f-8d48-8c77eb2bdd75","Type":"ContainerDied","Data":"a1b2f0987e9162c2b07f706325223d69cef9715ecc8c25e793e8743051442e88"} Nov 28 15:44:48 crc kubenswrapper[4817]: I1128 15:44:48.992574 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52hrt" event={"ID":"ab76693e-7ecc-474f-8d48-8c77eb2bdd75","Type":"ContainerStarted","Data":"973e35dfd546216e4354312fa8454dd27ffdb1e28b828e53689d6abd70a40592"} Nov 28 15:44:49 crc kubenswrapper[4817]: I1128 15:44:49.019501 4817 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-52hrt" podStartSLOduration=2.392131823 podStartE2EDuration="5.019476369s" podCreationTimestamp="2025-11-28 15:44:44 +0000 UTC" firstStartedPulling="2025-11-28 15:44:45.959415338 +0000 UTC m=+4508.547393604" lastFinishedPulling="2025-11-28 15:44:48.586759844 +0000 UTC m=+4511.174738150" observedRunningTime="2025-11-28 15:44:49.01391373 +0000 UTC m=+4511.601891986" watchObservedRunningTime="2025-11-28 15:44:49.019476369 +0000 UTC m=+4511.607454635" Nov 28 15:44:54 crc kubenswrapper[4817]: I1128 15:44:54.967335 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:54 crc kubenswrapper[4817]: I1128 15:44:54.967776 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:55 crc kubenswrapper[4817]: I1128 15:44:55.049089 4817 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:55 crc kubenswrapper[4817]: I1128 15:44:55.152783 4817 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:55 crc kubenswrapper[4817]: I1128 15:44:55.315188 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52hrt"] Nov 28 15:44:57 crc kubenswrapper[4817]: I1128 15:44:57.093278 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-52hrt" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="registry-server" containerID="cri-o://973e35dfd546216e4354312fa8454dd27ffdb1e28b828e53689d6abd70a40592" gracePeriod=2 Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.111262 4817 generic.go:334] "Generic (PLEG): container finished" podID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerID="973e35dfd546216e4354312fa8454dd27ffdb1e28b828e53689d6abd70a40592" exitCode=0 Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.111312 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52hrt" event={"ID":"ab76693e-7ecc-474f-8d48-8c77eb2bdd75","Type":"ContainerDied","Data":"973e35dfd546216e4354312fa8454dd27ffdb1e28b828e53689d6abd70a40592"} Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.737744 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:44:58 crc kubenswrapper[4817]: E1128 15:44:58.738310 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.821508 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.882582 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-catalog-content\") pod \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.882691 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x87ln\" (UniqueName: \"kubernetes.io/projected/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-kube-api-access-x87ln\") pod \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.882889 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-utilities\") pod \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\" (UID: \"ab76693e-7ecc-474f-8d48-8c77eb2bdd75\") " Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.884178 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-utilities" (OuterVolumeSpecName: "utilities") pod "ab76693e-7ecc-474f-8d48-8c77eb2bdd75" (UID: "ab76693e-7ecc-474f-8d48-8c77eb2bdd75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.885683 4817 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.897016 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-kube-api-access-x87ln" (OuterVolumeSpecName: "kube-api-access-x87ln") pod "ab76693e-7ecc-474f-8d48-8c77eb2bdd75" (UID: "ab76693e-7ecc-474f-8d48-8c77eb2bdd75"). InnerVolumeSpecName "kube-api-access-x87ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.904103 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab76693e-7ecc-474f-8d48-8c77eb2bdd75" (UID: "ab76693e-7ecc-474f-8d48-8c77eb2bdd75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.987446 4817 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:58 crc kubenswrapper[4817]: I1128 15:44:58.987480 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x87ln\" (UniqueName: \"kubernetes.io/projected/ab76693e-7ecc-474f-8d48-8c77eb2bdd75-kube-api-access-x87ln\") on node \"crc\" DevicePath \"\"" Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.131678 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-52hrt" event={"ID":"ab76693e-7ecc-474f-8d48-8c77eb2bdd75","Type":"ContainerDied","Data":"609dbdcbde830e23f1221efdd029afe5d627f7067c927404678fd14c0761bc64"} Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.131760 4817 scope.go:117] "RemoveContainer" containerID="973e35dfd546216e4354312fa8454dd27ffdb1e28b828e53689d6abd70a40592" Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.131775 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-52hrt" Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.176066 4817 scope.go:117] "RemoveContainer" containerID="a1b2f0987e9162c2b07f706325223d69cef9715ecc8c25e793e8743051442e88" Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.186335 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-52hrt"] Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.197877 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-52hrt"] Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.205860 4817 scope.go:117] "RemoveContainer" containerID="77e1849b262848d3716506efc6a44ef7ef34cae366e8e9fb889fb17904fc1c5d" Nov 28 15:44:59 crc kubenswrapper[4817]: I1128 15:44:59.756529 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" path="/var/lib/kubelet/pods/ab76693e-7ecc-474f-8d48-8c77eb2bdd75/volumes" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.202374 4817 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz"] Nov 28 15:45:00 crc kubenswrapper[4817]: E1128 15:45:00.203242 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="extract-content" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.203258 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="extract-content" Nov 28 15:45:00 crc kubenswrapper[4817]: E1128 15:45:00.203283 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="registry-server" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.203291 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="registry-server" Nov 28 15:45:00 crc kubenswrapper[4817]: E1128 15:45:00.203311 4817 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="extract-utilities" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.203320 4817 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="extract-utilities" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.203597 4817 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab76693e-7ecc-474f-8d48-8c77eb2bdd75" containerName="registry-server" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.204566 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.207339 4817 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.207606 4817 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.211982 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz"] Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.315218 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e278858-5423-47c3-b3f6-760c101cee31-secret-volume\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.315295 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4nmg\" (UniqueName: \"kubernetes.io/projected/1e278858-5423-47c3-b3f6-760c101cee31-kube-api-access-k4nmg\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.315511 4817 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e278858-5423-47c3-b3f6-760c101cee31-config-volume\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.417685 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e278858-5423-47c3-b3f6-760c101cee31-secret-volume\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.417814 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4nmg\" (UniqueName: \"kubernetes.io/projected/1e278858-5423-47c3-b3f6-760c101cee31-kube-api-access-k4nmg\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.417889 4817 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e278858-5423-47c3-b3f6-760c101cee31-config-volume\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.418660 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e278858-5423-47c3-b3f6-760c101cee31-config-volume\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.422501 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e278858-5423-47c3-b3f6-760c101cee31-secret-volume\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.434349 4817 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4nmg\" (UniqueName: \"kubernetes.io/projected/1e278858-5423-47c3-b3f6-760c101cee31-kube-api-access-k4nmg\") pod \"collect-profiles-29405745-pnnmz\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.527160 4817 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:00 crc kubenswrapper[4817]: I1128 15:45:00.967824 4817 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz"] Nov 28 15:45:00 crc kubenswrapper[4817]: W1128 15:45:00.985065 4817 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e278858_5423_47c3_b3f6_760c101cee31.slice/crio-b4bd1854c3fe1e28b974c32f973c5bc996c97385a7dab83b0745d4b6f12a039f WatchSource:0}: Error finding container b4bd1854c3fe1e28b974c32f973c5bc996c97385a7dab83b0745d4b6f12a039f: Status 404 returned error can't find the container with id b4bd1854c3fe1e28b974c32f973c5bc996c97385a7dab83b0745d4b6f12a039f Nov 28 15:45:01 crc kubenswrapper[4817]: I1128 15:45:01.173120 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" event={"ID":"1e278858-5423-47c3-b3f6-760c101cee31","Type":"ContainerStarted","Data":"b4bd1854c3fe1e28b974c32f973c5bc996c97385a7dab83b0745d4b6f12a039f"} Nov 28 15:45:02 crc kubenswrapper[4817]: I1128 15:45:02.183172 4817 generic.go:334] "Generic (PLEG): container finished" podID="1e278858-5423-47c3-b3f6-760c101cee31" containerID="d9b2c5bcbcac36d47a16f1b2cdc232108247fdb4fdc9840dbbec033d6be160d0" exitCode=0 Nov 28 15:45:02 crc kubenswrapper[4817]: I1128 15:45:02.183291 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" event={"ID":"1e278858-5423-47c3-b3f6-760c101cee31","Type":"ContainerDied","Data":"d9b2c5bcbcac36d47a16f1b2cdc232108247fdb4fdc9840dbbec033d6be160d0"} Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.555819 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.577563 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e278858-5423-47c3-b3f6-760c101cee31-config-volume\") pod \"1e278858-5423-47c3-b3f6-760c101cee31\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.577636 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4nmg\" (UniqueName: \"kubernetes.io/projected/1e278858-5423-47c3-b3f6-760c101cee31-kube-api-access-k4nmg\") pod \"1e278858-5423-47c3-b3f6-760c101cee31\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.577794 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e278858-5423-47c3-b3f6-760c101cee31-secret-volume\") pod \"1e278858-5423-47c3-b3f6-760c101cee31\" (UID: \"1e278858-5423-47c3-b3f6-760c101cee31\") " Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.578459 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e278858-5423-47c3-b3f6-760c101cee31-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e278858-5423-47c3-b3f6-760c101cee31" (UID: "1e278858-5423-47c3-b3f6-760c101cee31"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.592813 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e278858-5423-47c3-b3f6-760c101cee31-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e278858-5423-47c3-b3f6-760c101cee31" (UID: "1e278858-5423-47c3-b3f6-760c101cee31"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.593326 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e278858-5423-47c3-b3f6-760c101cee31-kube-api-access-k4nmg" (OuterVolumeSpecName: "kube-api-access-k4nmg") pod "1e278858-5423-47c3-b3f6-760c101cee31" (UID: "1e278858-5423-47c3-b3f6-760c101cee31"). InnerVolumeSpecName "kube-api-access-k4nmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.680466 4817 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e278858-5423-47c3-b3f6-760c101cee31-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.680528 4817 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e278858-5423-47c3-b3f6-760c101cee31-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:03 crc kubenswrapper[4817]: I1128 15:45:03.680544 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4nmg\" (UniqueName: \"kubernetes.io/projected/1e278858-5423-47c3-b3f6-760c101cee31-kube-api-access-k4nmg\") on node \"crc\" DevicePath \"\"" Nov 28 15:45:04 crc kubenswrapper[4817]: I1128 15:45:04.208157 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" event={"ID":"1e278858-5423-47c3-b3f6-760c101cee31","Type":"ContainerDied","Data":"b4bd1854c3fe1e28b974c32f973c5bc996c97385a7dab83b0745d4b6f12a039f"} Nov 28 15:45:04 crc kubenswrapper[4817]: I1128 15:45:04.208517 4817 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4bd1854c3fe1e28b974c32f973c5bc996c97385a7dab83b0745d4b6f12a039f" Nov 28 15:45:04 crc kubenswrapper[4817]: I1128 15:45:04.208270 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405745-pnnmz" Nov 28 15:45:04 crc kubenswrapper[4817]: I1128 15:45:04.652261 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp"] Nov 28 15:45:04 crc kubenswrapper[4817]: I1128 15:45:04.663543 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405700-gzgbp"] Nov 28 15:45:05 crc kubenswrapper[4817]: I1128 15:45:05.752957 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10b209ee-9bd3-4e8b-a1a1-f74e08a431f6" path="/var/lib/kubelet/pods/10b209ee-9bd3-4e8b-a1a1-f74e08a431f6/volumes" Nov 28 15:45:10 crc kubenswrapper[4817]: I1128 15:45:10.736792 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:45:10 crc kubenswrapper[4817]: E1128 15:45:10.737657 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:45:16 crc kubenswrapper[4817]: I1128 15:45:16.131969 4817 scope.go:117] "RemoveContainer" containerID="120f5c697586aeaa59bcd06caeee7c552acae7a9ab3422fb9a50a7afb89a0477" Nov 28 15:45:24 crc kubenswrapper[4817]: I1128 15:45:24.737150 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:45:24 crc kubenswrapper[4817]: E1128 15:45:24.738157 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:45:38 crc kubenswrapper[4817]: I1128 15:45:38.737653 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:45:38 crc kubenswrapper[4817]: E1128 15:45:38.738551 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:45:52 crc kubenswrapper[4817]: I1128 15:45:52.735292 4817 generic.go:334] "Generic (PLEG): container finished" podID="56977377-a544-492d-803a-a98943ea22fc" containerID="e3cf04c6b2e68290d49e2c14a573ee03261b67aec99fa89dc1c1b92c83b3a72e" exitCode=0 Nov 28 15:45:52 crc kubenswrapper[4817]: I1128 15:45:52.735393 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9q2hs/must-gather-vcl74" event={"ID":"56977377-a544-492d-803a-a98943ea22fc","Type":"ContainerDied","Data":"e3cf04c6b2e68290d49e2c14a573ee03261b67aec99fa89dc1c1b92c83b3a72e"} Nov 28 15:45:52 crc kubenswrapper[4817]: I1128 15:45:52.736664 4817 scope.go:117] "RemoveContainer" containerID="e3cf04c6b2e68290d49e2c14a573ee03261b67aec99fa89dc1c1b92c83b3a72e" Nov 28 15:45:52 crc kubenswrapper[4817]: I1128 15:45:52.843306 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9q2hs_must-gather-vcl74_56977377-a544-492d-803a-a98943ea22fc/gather/0.log" Nov 28 15:45:53 crc kubenswrapper[4817]: I1128 15:45:53.739939 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:45:53 crc kubenswrapper[4817]: E1128 15:45:53.743828 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:46:03 crc kubenswrapper[4817]: I1128 15:46:03.550839 4817 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9q2hs/must-gather-vcl74"] Nov 28 15:46:03 crc kubenswrapper[4817]: I1128 15:46:03.551643 4817 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9q2hs/must-gather-vcl74" podUID="56977377-a544-492d-803a-a98943ea22fc" containerName="copy" containerID="cri-o://aa6054d5561cd4e3d571a5b17d9e8edc81cfe037dd28257d70118633d5a439c6" gracePeriod=2 Nov 28 15:46:03 crc kubenswrapper[4817]: I1128 15:46:03.565612 4817 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9q2hs/must-gather-vcl74"] Nov 28 15:46:03 crc kubenswrapper[4817]: I1128 15:46:03.836286 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9q2hs_must-gather-vcl74_56977377-a544-492d-803a-a98943ea22fc/copy/0.log" Nov 28 15:46:03 crc kubenswrapper[4817]: I1128 15:46:03.836628 4817 generic.go:334] "Generic (PLEG): container finished" podID="56977377-a544-492d-803a-a98943ea22fc" containerID="aa6054d5561cd4e3d571a5b17d9e8edc81cfe037dd28257d70118633d5a439c6" exitCode=143 Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.027450 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9q2hs_must-gather-vcl74_56977377-a544-492d-803a-a98943ea22fc/copy/0.log" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.028167 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.154414 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/56977377-a544-492d-803a-a98943ea22fc-must-gather-output\") pod \"56977377-a544-492d-803a-a98943ea22fc\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.154635 4817 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjjz4\" (UniqueName: \"kubernetes.io/projected/56977377-a544-492d-803a-a98943ea22fc-kube-api-access-pjjz4\") pod \"56977377-a544-492d-803a-a98943ea22fc\" (UID: \"56977377-a544-492d-803a-a98943ea22fc\") " Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.161010 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56977377-a544-492d-803a-a98943ea22fc-kube-api-access-pjjz4" (OuterVolumeSpecName: "kube-api-access-pjjz4") pod "56977377-a544-492d-803a-a98943ea22fc" (UID: "56977377-a544-492d-803a-a98943ea22fc"). InnerVolumeSpecName "kube-api-access-pjjz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.256694 4817 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjjz4\" (UniqueName: \"kubernetes.io/projected/56977377-a544-492d-803a-a98943ea22fc-kube-api-access-pjjz4\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.322567 4817 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56977377-a544-492d-803a-a98943ea22fc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "56977377-a544-492d-803a-a98943ea22fc" (UID: "56977377-a544-492d-803a-a98943ea22fc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.359164 4817 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/56977377-a544-492d-803a-a98943ea22fc-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.844614 4817 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9q2hs_must-gather-vcl74_56977377-a544-492d-803a-a98943ea22fc/copy/0.log" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.845124 4817 scope.go:117] "RemoveContainer" containerID="aa6054d5561cd4e3d571a5b17d9e8edc81cfe037dd28257d70118633d5a439c6" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.845156 4817 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9q2hs/must-gather-vcl74" Nov 28 15:46:04 crc kubenswrapper[4817]: I1128 15:46:04.864233 4817 scope.go:117] "RemoveContainer" containerID="e3cf04c6b2e68290d49e2c14a573ee03261b67aec99fa89dc1c1b92c83b3a72e" Nov 28 15:46:05 crc kubenswrapper[4817]: I1128 15:46:05.736671 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:46:05 crc kubenswrapper[4817]: E1128 15:46:05.737209 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:46:05 crc kubenswrapper[4817]: I1128 15:46:05.746833 4817 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56977377-a544-492d-803a-a98943ea22fc" path="/var/lib/kubelet/pods/56977377-a544-492d-803a-a98943ea22fc/volumes" Nov 28 15:46:20 crc kubenswrapper[4817]: I1128 15:46:20.737277 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:46:20 crc kubenswrapper[4817]: E1128 15:46:20.738137 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:46:32 crc kubenswrapper[4817]: I1128 15:46:32.737514 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:46:32 crc kubenswrapper[4817]: E1128 15:46:32.738330 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:46:44 crc kubenswrapper[4817]: I1128 15:46:44.736863 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:46:44 crc kubenswrapper[4817]: E1128 15:46:44.737606 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:46:58 crc kubenswrapper[4817]: I1128 15:46:58.737816 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:46:58 crc kubenswrapper[4817]: E1128 15:46:58.738802 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:47:09 crc kubenswrapper[4817]: I1128 15:47:09.737953 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:47:09 crc kubenswrapper[4817]: E1128 15:47:09.738788 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:47:23 crc kubenswrapper[4817]: I1128 15:47:23.737291 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:47:23 crc kubenswrapper[4817]: E1128 15:47:23.740424 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:47:37 crc kubenswrapper[4817]: I1128 15:47:37.748626 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:47:37 crc kubenswrapper[4817]: E1128 15:47:37.752438 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:47:48 crc kubenswrapper[4817]: I1128 15:47:48.738045 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:47:48 crc kubenswrapper[4817]: E1128 15:47:48.739389 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:47:59 crc kubenswrapper[4817]: I1128 15:47:59.737251 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:47:59 crc kubenswrapper[4817]: E1128 15:47:59.738005 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:48:12 crc kubenswrapper[4817]: I1128 15:48:12.737615 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:48:12 crc kubenswrapper[4817]: E1128 15:48:12.738810 4817 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6kh49_openshift-machine-config-operator(2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc)\"" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" podUID="2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc" Nov 28 15:48:25 crc kubenswrapper[4817]: I1128 15:48:25.737255 4817 scope.go:117] "RemoveContainer" containerID="683f93078c351563d98ee5f4ad32afe2d5669c67d3d70a5ee07db7ef8c9f2a41" Nov 28 15:48:26 crc kubenswrapper[4817]: I1128 15:48:26.360517 4817 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6kh49" event={"ID":"2f2c731e-6fa9-4e0c-81dd-115f13fa4dfc","Type":"ContainerStarted","Data":"c02b198e887a42bca6cb3eb66d169badad325be502f1f1811b8f25fe6ba98559"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112342203024436 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112342203017353 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112330372016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112330372015452 5ustar corecore